diff --git "a/sg/antmaze-medium-diverse-v0/1/debug.log" "b/sg/antmaze-medium-diverse-v0/1/debug.log" new file mode 100644--- /dev/null +++ "b/sg/antmaze-medium-diverse-v0/1/debug.log" @@ -0,0 +1,52000 @@ +2022-05-10 13:11:21.553143 PDT | [1] Epoch -1000 finished +---------------------------------- ---------------- +epoch -1000 +replay_buffer/size 999033 +trainer/num train calls 1000 +trainer/Policy Loss 38.6967 +trainer/Log Pis Mean -5.34101 +trainer/Log Pis Std 0.648232 +trainer/Log Pis Max -3.39717 +trainer/Log Pis Min -7.49236 +trainer/policy/mean Mean 5.72749e-05 +trainer/policy/mean Std 0.00020513 +trainer/policy/mean Max 0.000794743 +trainer/policy/mean Min -0.00066434 +trainer/policy/normal/std Mean 1.00017 +trainer/policy/normal/std Std 0.000709215 +trainer/policy/normal/std Max 1.00131 +trainer/policy/normal/std Min 0.998663 +trainer/policy/normal/log_std Mean 0.000168378 +trainer/policy/normal/log_std Std 0.000709184 +trainer/policy/normal/log_std Max 0.00130866 +trainer/policy/normal/log_std Min -0.00133765 +eval/num steps total 1000 +eval/num paths total 1 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 3.51531e-05 +eval/Actions Std 0.000123447 +eval/Actions Max 0.000534459 +eval/Actions Min -0.000451015 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 4.7424 +time/logging (s) 0.00753811 +time/sampling batch (s) 0.543318 +time/saving (s) 0.00630517 +time/training (s) 7.70157 +time/epoch (s) 13.0011 +time/total (s) 48.1223 +Epoch -1000 +---------------------------------- ---------------- +2022-05-10 13:11:33.395416 PDT | [1] Epoch -999 finished +---------------------------------- --------------- +epoch -999 +replay_buffer/size 999033 +trainer/num train calls 2000 +trainer/Policy Loss -16.5107 +trainer/Log Pis Mean 22.0592 +trainer/Log Pis Std 14.1017 +trainer/Log Pis Max 88.2973 +trainer/Log Pis Min -7.11501 +trainer/policy/mean Mean -0.0424403 +trainer/policy/mean Std 0.719386 +trainer/policy/mean Max 0.992759 +trainer/policy/mean Min -0.994114 +trainer/policy/normal/std Mean 3.94535 +trainer/policy/normal/std Std 0.553957 +trainer/policy/normal/std Max 7.33796 +trainer/policy/normal/std Min 2.58814 +trainer/policy/normal/log_std Mean 1.36306 +trainer/policy/normal/log_std Std 0.136842 +trainer/policy/normal/log_std Max 1.99306 +trainer/policy/normal/log_std Min 0.950939 +eval/num steps total 2000 +eval/num paths total 2 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00182786 +eval/Actions Std 0.705267 +eval/Actions Max 0.978889 +eval/Actions Min -0.982175 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73459 +time/logging (s) 0.00397019 +time/sampling batch (s) 0.279985 +time/saving (s) 0.00381367 +time/training (s) 8.79355 +time/epoch (s) 11.8159 +time/total (s) 59.9438 +Epoch -999 +---------------------------------- --------------- +2022-05-10 13:11:43.911276 PDT | [1] Epoch -998 finished +---------------------------------- --------------- +epoch -998 +replay_buffer/size 999033 +trainer/num train calls 3000 +trainer/Policy Loss -18.4132 +trainer/Log Pis Mean 22.5787 +trainer/Log Pis Std 13.2036 +trainer/Log Pis Max 63.0687 +trainer/Log Pis Min -10.2818 +trainer/policy/mean Mean -0.0604008 +trainer/policy/mean Std 0.830975 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999965 +trainer/policy/normal/std Mean 3.51924 +trainer/policy/normal/std Std 0.491062 +trainer/policy/normal/std Max 5.75603 +trainer/policy/normal/std Min 1.89373 +trainer/policy/normal/log_std Mean 1.24828 +trainer/policy/normal/log_std Std 0.142508 +trainer/policy/normal/log_std Max 1.75025 +trainer/policy/normal/log_std Min 0.638549 +eval/num steps total 3000 +eval/num paths total 3 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0310601 +eval/Actions Std 0.676248 +eval/Actions Max 0.998125 +eval/Actions Min -0.982032 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47902 +time/logging (s) 0.00422196 +time/sampling batch (s) 0.277258 +time/saving (s) 0.00406811 +time/training (s) 7.73199 +time/epoch (s) 10.4966 +time/total (s) 70.4436 +Epoch -998 +---------------------------------- --------------- +2022-05-10 13:11:53.634513 PDT | [1] Epoch -997 finished +---------------------------------- --------------- +epoch -997 +replay_buffer/size 999033 +trainer/num train calls 4000 +trainer/Policy Loss -16.8917 +trainer/Log Pis Mean 22.5308 +trainer/Log Pis Std 12.8631 +trainer/Log Pis Max 62.5833 +trainer/Log Pis Min -6.48641 +trainer/policy/mean Mean -0.0440601 +trainer/policy/mean Std 0.854449 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 3.34103 +trainer/policy/normal/std Std 0.461096 +trainer/policy/normal/std Max 5.24411 +trainer/policy/normal/std Min 1.55138 +trainer/policy/normal/log_std Mean 1.19634 +trainer/policy/normal/log_std Std 0.143387 +trainer/policy/normal/log_std Max 1.65711 +trainer/policy/normal/log_std Min 0.439142 +eval/num steps total 4000 +eval/num paths total 4 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0549362 +eval/Actions Std 0.80416 +eval/Actions Max 0.999496 +eval/Actions Min -0.998804 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50664 +time/logging (s) 0.00366903 +time/sampling batch (s) 0.27739 +time/saving (s) 0.00344849 +time/training (s) 6.91198 +time/epoch (s) 9.70312 +time/total (s) 80.1501 +Epoch -997 +---------------------------------- --------------- +2022-05-10 13:12:03.744601 PDT | [1] Epoch -996 finished +---------------------------------- --------------- +epoch -996 +replay_buffer/size 999033 +trainer/num train calls 5000 +trainer/Policy Loss -17.863 +trainer/Log Pis Mean 23.0013 +trainer/Log Pis Std 12.3677 +trainer/Log Pis Max 71.6084 +trainer/Log Pis Min -11.0342 +trainer/policy/mean Mean -0.0270586 +trainer/policy/mean Std 0.879765 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 3.20909 +trainer/policy/normal/std Std 0.43076 +trainer/policy/normal/std Max 5.46024 +trainer/policy/normal/std Min 1.49982 +trainer/policy/normal/log_std Mean 1.15665 +trainer/policy/normal/log_std Std 0.138961 +trainer/policy/normal/log_std Max 1.69749 +trainer/policy/normal/log_std Min 0.405343 +eval/num steps total 5000 +eval/num paths total 5 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0939369 +eval/Actions Std 0.833461 +eval/Actions Max 0.999995 +eval/Actions Min -0.999972 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7632 +time/logging (s) 0.00412606 +time/sampling batch (s) 0.282009 +time/saving (s) 0.00399362 +time/training (s) 7.03771 +time/epoch (s) 10.091 +time/total (s) 90.244 +Epoch -996 +---------------------------------- --------------- +2022-05-10 13:12:13.507077 PDT | [1] Epoch -995 finished +---------------------------------- --------------- +epoch -995 +replay_buffer/size 999033 +trainer/num train calls 6000 +trainer/Policy Loss -18.5126 +trainer/Log Pis Mean 22.9997 +trainer/Log Pis Std 12.867 +trainer/Log Pis Max 69.6586 +trainer/Log Pis Min -5.76067 +trainer/policy/mean Mean -0.0262533 +trainer/policy/mean Std 0.883386 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 3.17686 +trainer/policy/normal/std Std 0.425123 +trainer/policy/normal/std Max 5.04988 +trainer/policy/normal/std Min 1.35154 +trainer/policy/normal/log_std Mean 1.14618 +trainer/policy/normal/log_std Std 0.143336 +trainer/policy/normal/log_std Max 1.61936 +trainer/policy/normal/log_std Min 0.301246 +eval/num steps total 6000 +eval/num paths total 6 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.13088 +eval/Actions Std 0.829851 +eval/Actions Max 0.999964 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68369 +time/logging (s) 0.00379248 +time/sampling batch (s) 0.283219 +time/saving (s) 0.00359554 +time/training (s) 6.76765 +time/epoch (s) 9.74196 +time/total (s) 99.9893 +Epoch -995 +---------------------------------- --------------- +2022-05-10 13:12:23.894085 PDT | [1] Epoch -994 finished +---------------------------------- ---------------- +epoch -994 +replay_buffer/size 999033 +trainer/num train calls 7000 +trainer/Policy Loss -18.6163 +trainer/Log Pis Mean 23.8661 +trainer/Log Pis Std 13.4004 +trainer/Log Pis Max 59.0391 +trainer/Log Pis Min -7.47314 +trainer/policy/mean Mean -0.0342965 +trainer/policy/mean Std 0.894212 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.11501 +trainer/policy/normal/std Std 0.413459 +trainer/policy/normal/std Max 5.68707 +trainer/policy/normal/std Min 1.25569 +trainer/policy/normal/log_std Mean 1.1267 +trainer/policy/normal/log_std Std 0.141956 +trainer/policy/normal/log_std Max 1.7382 +trainer/policy/normal/log_std Min 0.227689 +eval/num steps total 7000 +eval/num paths total 7 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000255507 +eval/Actions Std 0.868947 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60577 +time/logging (s) 0.00361967 +time/sampling batch (s) 0.286054 +time/saving (s) 0.00341581 +time/training (s) 7.46818 +time/epoch (s) 10.367 +time/total (s) 110.359 +Epoch -994 +---------------------------------- ---------------- +2022-05-10 13:12:34.128047 PDT | [1] Epoch -993 finished +---------------------------------- --------------- +epoch -993 +replay_buffer/size 999033 +trainer/num train calls 8000 +trainer/Policy Loss -17.0912 +trainer/Log Pis Mean 24.149 +trainer/Log Pis Std 13.1325 +trainer/Log Pis Max 69.8912 +trainer/Log Pis Min -8.13968 +trainer/policy/mean Mean -0.031964 +trainer/policy/mean Std 0.884175 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.13268 +trainer/policy/normal/std Std 0.417089 +trainer/policy/normal/std Max 5.0571 +trainer/policy/normal/std Min 1.2115 +trainer/policy/normal/log_std Mean 1.13191 +trainer/policy/normal/log_std Std 0.146839 +trainer/policy/normal/log_std Max 1.62079 +trainer/policy/normal/log_std Min 0.191861 +eval/num steps total 8000 +eval/num paths total 8 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138189 +eval/Actions Std 0.887491 +eval/Actions Max 0.999993 +eval/Actions Min -0.999959 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79543 +time/logging (s) 0.0036738 +time/sampling batch (s) 0.279648 +time/saving (s) 0.00338737 +time/training (s) 7.13245 +time/epoch (s) 10.2146 +time/total (s) 120.577 +Epoch -993 +---------------------------------- --------------- +2022-05-10 13:12:45.670554 PDT | [1] Epoch -992 finished +---------------------------------- --------------- +epoch -992 +replay_buffer/size 999033 +trainer/num train calls 9000 +trainer/Policy Loss -19.3675 +trainer/Log Pis Mean 23.591 +trainer/Log Pis Std 12.7853 +trainer/Log Pis Max 74.4086 +trainer/Log Pis Min -5.02655 +trainer/policy/mean Mean -0.0226047 +trainer/policy/mean Std 0.889169 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.14694 +trainer/policy/normal/std Std 0.434605 +trainer/policy/normal/std Max 5.19725 +trainer/policy/normal/std Min 1.12895 +trainer/policy/normal/log_std Mean 1.13577 +trainer/policy/normal/log_std Std 0.151772 +trainer/policy/normal/log_std Max 1.64813 +trainer/policy/normal/log_std Min 0.12129 +eval/num steps total 9000 +eval/num paths total 9 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121407 +eval/Actions Std 0.873711 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75658 +time/logging (s) 0.00370734 +time/sampling batch (s) 0.292582 +time/saving (s) 0.00342868 +time/training (s) 8.46625 +time/epoch (s) 11.5225 +time/total (s) 132.102 +Epoch -992 +---------------------------------- --------------- +2022-05-10 13:12:55.932259 PDT | [1] Epoch -991 finished +---------------------------------- --------------- +epoch -991 +replay_buffer/size 999033 +trainer/num train calls 10000 +trainer/Policy Loss -18.4437 +trainer/Log Pis Mean 23.3695 +trainer/Log Pis Std 12.5225 +trainer/Log Pis Max 69.459 +trainer/Log Pis Min -5.7264 +trainer/policy/mean Mean -0.0400293 +trainer/policy/mean Std 0.890607 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.10315 +trainer/policy/normal/std Std 0.437811 +trainer/policy/normal/std Max 6.38145 +trainer/policy/normal/std Min 1.0549 +trainer/policy/normal/log_std Mean 1.12109 +trainer/policy/normal/log_std Std 0.157473 +trainer/policy/normal/log_std Max 1.8534 +trainer/policy/normal/log_std Min 0.0534422 +eval/num steps total 10000 +eval/num paths total 10 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0935633 +eval/Actions Std 0.861729 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70316 +time/logging (s) 0.00420028 +time/sampling batch (s) 0.280149 +time/saving (s) 0.00379133 +time/training (s) 7.25143 +time/epoch (s) 10.2427 +time/total (s) 142.348 +Epoch -991 +---------------------------------- --------------- +2022-05-10 13:13:05.393310 PDT | [1] Epoch -990 finished +---------------------------------- --------------- +epoch -990 +replay_buffer/size 999033 +trainer/num train calls 11000 +trainer/Policy Loss -18.885 +trainer/Log Pis Mean 22.8935 +trainer/Log Pis Std 12.8925 +trainer/Log Pis Max 63.9677 +trainer/Log Pis Min -7.22985 +trainer/policy/mean Mean -0.029093 +trainer/policy/mean Std 0.889433 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 3.05686 +trainer/policy/normal/std Std 0.420428 +trainer/policy/normal/std Max 4.62666 +trainer/policy/normal/std Min 1.03207 +trainer/policy/normal/log_std Mean 1.10625 +trainer/policy/normal/log_std Std 0.157315 +trainer/policy/normal/log_std Max 1.53184 +trainer/policy/normal/log_std Min 0.0315643 +eval/num steps total 11000 +eval/num paths total 11 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196798 +eval/Actions Std 0.93578 +eval/Actions Max 0.999996 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4602 +time/logging (s) 0.00387744 +time/sampling batch (s) 0.779327 +time/saving (s) 0.00367476 +time/training (s) 6.19384 +time/epoch (s) 9.44092 +time/total (s) 151.792 +Epoch -990 +---------------------------------- --------------- +2022-05-10 13:13:14.937779 PDT | [1] Epoch -989 finished +---------------------------------- --------------- +epoch -989 +replay_buffer/size 999033 +trainer/num train calls 12000 +trainer/Policy Loss -20.0722 +trainer/Log Pis Mean 23.7828 +trainer/Log Pis Std 12.7112 +trainer/Log Pis Max 63.7978 +trainer/Log Pis Min -11.2209 +trainer/policy/mean Mean -0.0291343 +trainer/policy/mean Std 0.891973 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.99248 +trainer/policy/normal/std Std 0.434641 +trainer/policy/normal/std Max 4.89364 +trainer/policy/normal/std Min 1.03631 +trainer/policy/normal/log_std Mean 1.08389 +trainer/policy/normal/log_std Std 0.163963 +trainer/policy/normal/log_std Max 1.58794 +trainer/policy/normal/log_std Min 0.035668 +eval/num steps total 12000 +eval/num paths total 12 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103179 +eval/Actions Std 0.886491 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53535 +time/logging (s) 0.00387643 +time/sampling batch (s) 0.530546 +time/saving (s) 0.00347026 +time/training (s) 6.45166 +time/epoch (s) 9.5249 +time/total (s) 161.32 +Epoch -989 +---------------------------------- --------------- +2022-05-10 13:13:24.816229 PDT | [1] Epoch -988 finished +---------------------------------- --------------- +epoch -988 +replay_buffer/size 999033 +trainer/num train calls 13000 +trainer/Policy Loss -19.2031 +trainer/Log Pis Mean 24.5207 +trainer/Log Pis Std 13.8346 +trainer/Log Pis Max 75.6213 +trainer/Log Pis Min -3.93741 +trainer/policy/mean Mean -0.033677 +trainer/policy/mean Std 0.89167 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 3.07651 +trainer/policy/normal/std Std 0.451068 +trainer/policy/normal/std Max 4.97874 +trainer/policy/normal/std Min 0.919117 +trainer/policy/normal/log_std Mean 1.11115 +trainer/policy/normal/log_std Std 0.167703 +trainer/policy/normal/log_std Max 1.60518 +trainer/policy/normal/log_std Min -0.0843419 +eval/num steps total 13000 +eval/num paths total 13 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.347719 +eval/Actions Std 0.88491 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39221 +time/logging (s) 0.00377053 +time/sampling batch (s) 0.542893 +time/saving (s) 0.00364908 +time/training (s) 6.91618 +time/epoch (s) 9.85871 +time/total (s) 171.181 +Epoch -988 +---------------------------------- --------------- +2022-05-10 13:13:35.605198 PDT | [1] Epoch -987 finished +---------------------------------- --------------- +epoch -987 +replay_buffer/size 999033 +trainer/num train calls 14000 +trainer/Policy Loss -19.3893 +trainer/Log Pis Mean 23.3791 +trainer/Log Pis Std 12.6224 +trainer/Log Pis Max 82.5196 +trainer/Log Pis Min -6.73677 +trainer/policy/mean Mean -0.0190203 +trainer/policy/mean Std 0.888037 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 3.03712 +trainer/policy/normal/std Std 0.438769 +trainer/policy/normal/std Max 4.65485 +trainer/policy/normal/std Min 0.97908 +trainer/policy/normal/log_std Mean 1.0984 +trainer/policy/normal/log_std Std 0.167475 +trainer/policy/normal/log_std Max 1.53791 +trainer/policy/normal/log_std Min -0.0211418 +eval/num steps total 14000 +eval/num paths total 14 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.252455 +eval/Actions Std 0.91935 +eval/Actions Max 0.999996 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86706 +time/logging (s) 0.00392077 +time/sampling batch (s) 0.548748 +time/saving (s) 0.00364777 +time/training (s) 7.34594 +time/epoch (s) 10.7693 +time/total (s) 181.954 +Epoch -987 +---------------------------------- --------------- +2022-05-10 13:13:45.796554 PDT | [1] Epoch -986 finished +---------------------------------- --------------- +epoch -986 +replay_buffer/size 999033 +trainer/num train calls 15000 +trainer/Policy Loss -18.714 +trainer/Log Pis Mean 24.763 +trainer/Log Pis Std 13.642 +trainer/Log Pis Max 69.6295 +trainer/Log Pis Min -5.0374 +trainer/policy/mean Mean -0.0161627 +trainer/policy/mean Std 0.893383 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 3.04507 +trainer/policy/normal/std Std 0.432011 +trainer/policy/normal/std Max 4.6484 +trainer/policy/normal/std Min 0.912396 +trainer/policy/normal/log_std Mean 1.10151 +trainer/policy/normal/log_std Std 0.164022 +trainer/policy/normal/log_std Max 1.53652 +trainer/policy/normal/log_std Min -0.0916809 +eval/num steps total 15000 +eval/num paths total 15 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.029141 +eval/Actions Std 0.86833 +eval/Actions Max 0.999998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30671 +time/logging (s) 0.00388035 +time/sampling batch (s) 0.531807 +time/saving (s) 0.00344511 +time/training (s) 7.32569 +time/epoch (s) 10.1715 +time/total (s) 192.128 +Epoch -986 +---------------------------------- --------------- +2022-05-10 13:13:56.161799 PDT | [1] Epoch -985 finished +---------------------------------- --------------- +epoch -985 +replay_buffer/size 999033 +trainer/num train calls 16000 +trainer/Policy Loss -19.3838 +trainer/Log Pis Mean 23.0067 +trainer/Log Pis Std 12.8838 +trainer/Log Pis Max 69.7614 +trainer/Log Pis Min -9.46912 +trainer/policy/mean Mean -0.0223933 +trainer/policy/mean Std 0.894776 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.97147 +trainer/policy/normal/std Std 0.459191 +trainer/policy/normal/std Max 4.75768 +trainer/policy/normal/std Min 0.824974 +trainer/policy/normal/log_std Mean 1.0744 +trainer/policy/normal/log_std Std 0.182707 +trainer/policy/normal/log_std Max 1.55976 +trainer/policy/normal/log_std Min -0.192403 +eval/num steps total 16000 +eval/num paths total 16 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10463 +eval/Actions Std 0.895002 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66015 +time/logging (s) 0.0040015 +time/sampling batch (s) 0.53173 +time/saving (s) 0.00397044 +time/training (s) 7.14594 +time/epoch (s) 10.3458 +time/total (s) 202.477 +Epoch -985 +---------------------------------- --------------- +2022-05-10 13:14:05.602098 PDT | [1] Epoch -984 finished +---------------------------------- --------------- +epoch -984 +replay_buffer/size 999033 +trainer/num train calls 17000 +trainer/Policy Loss -18.9361 +trainer/Log Pis Mean 24.859 +trainer/Log Pis Std 12.4458 +trainer/Log Pis Max 62.7891 +trainer/Log Pis Min -8.38823 +trainer/policy/mean Mean -0.0468938 +trainer/policy/mean Std 0.897191 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.01295 +trainer/policy/normal/std Std 0.460221 +trainer/policy/normal/std Max 4.94949 +trainer/policy/normal/std Min 0.910013 +trainer/policy/normal/log_std Mean 1.08851 +trainer/policy/normal/log_std Std 0.181466 +trainer/policy/normal/log_std Max 1.59928 +trainer/policy/normal/log_std Min -0.0942965 +eval/num steps total 17000 +eval/num paths total 17 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196846 +eval/Actions Std 0.903974 +eval/Actions Max 0.999979 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69362 +time/logging (s) 0.00439979 +time/sampling batch (s) 0.28174 +time/saving (s) 0.00349834 +time/training (s) 6.43744 +time/epoch (s) 9.4207 +time/total (s) 211.901 +Epoch -984 +---------------------------------- --------------- +2022-05-10 13:14:15.213575 PDT | [1] Epoch -983 finished +---------------------------------- --------------- +epoch -983 +replay_buffer/size 999033 +trainer/num train calls 18000 +trainer/Policy Loss -18.563 +trainer/Log Pis Mean 23.9514 +trainer/Log Pis Std 13.37 +trainer/Log Pis Max 72.2514 +trainer/Log Pis Min -13.9109 +trainer/policy/mean Mean -0.050015 +trainer/policy/mean Std 0.887317 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 3.00851 +trainer/policy/normal/std Std 0.471921 +trainer/policy/normal/std Max 5.94426 +trainer/policy/normal/std Min 0.790714 +trainer/policy/normal/log_std Mean 1.08682 +trainer/policy/normal/log_std Std 0.181069 +trainer/policy/normal/log_std Max 1.78243 +trainer/policy/normal/log_std Min -0.234819 +eval/num steps total 18000 +eval/num paths total 18 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105401 +eval/Actions Std 0.900273 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57242 +time/logging (s) 0.0037659 +time/sampling batch (s) 0.282997 +time/saving (s) 0.00343536 +time/training (s) 6.72828 +time/epoch (s) 9.5909 +time/total (s) 221.495 +Epoch -983 +---------------------------------- --------------- +2022-05-10 13:14:25.909385 PDT | [1] Epoch -982 finished +---------------------------------- --------------- +epoch -982 +replay_buffer/size 999033 +trainer/num train calls 19000 +trainer/Policy Loss -18.1687 +trainer/Log Pis Mean 24.6069 +trainer/Log Pis Std 13.1732 +trainer/Log Pis Max 67.179 +trainer/Log Pis Min -6.82343 +trainer/policy/mean Mean -0.0429469 +trainer/policy/mean Std 0.894662 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 3.03945 +trainer/policy/normal/std Std 0.47006 +trainer/policy/normal/std Max 5.97457 +trainer/policy/normal/std Min 0.892177 +trainer/policy/normal/log_std Mean 1.09722 +trainer/policy/normal/log_std Std 0.180782 +trainer/policy/normal/log_std Max 1.78751 +trainer/policy/normal/log_std Min -0.114091 +eval/num steps total 19000 +eval/num paths total 19 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102417 +eval/Actions Std 0.888593 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67429 +time/logging (s) 0.00374943 +time/sampling batch (s) 0.533857 +time/saving (s) 0.00347722 +time/training (s) 7.46078 +time/epoch (s) 10.6762 +time/total (s) 232.174 +Epoch -982 +---------------------------------- --------------- +2022-05-10 13:14:36.464520 PDT | [1] Epoch -981 finished +---------------------------------- --------------- +epoch -981 +replay_buffer/size 999033 +trainer/num train calls 20000 +trainer/Policy Loss -19.1436 +trainer/Log Pis Mean 22.9583 +trainer/Log Pis Std 13.2603 +trainer/Log Pis Max 63.8574 +trainer/Log Pis Min -7.27686 +trainer/policy/mean Mean -0.026006 +trainer/policy/mean Std 0.900192 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.92194 +trainer/policy/normal/std Std 0.472047 +trainer/policy/normal/std Max 5.12396 +trainer/policy/normal/std Min 0.836565 +trainer/policy/normal/log_std Mean 1.05635 +trainer/policy/normal/log_std Std 0.189846 +trainer/policy/normal/log_std Max 1.63393 +trainer/policy/normal/log_std Min -0.178451 +eval/num steps total 20000 +eval/num paths total 20 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.198043 +eval/Actions Std 0.859719 +eval/Actions Max 1 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79052 +time/logging (s) 0.00369016 +time/sampling batch (s) 0.533116 +time/saving (s) 0.00341291 +time/training (s) 7.20445 +time/epoch (s) 10.5352 +time/total (s) 242.712 +Epoch -981 +---------------------------------- --------------- +2022-05-10 13:14:46.213584 PDT | [1] Epoch -980 finished +---------------------------------- --------------- +epoch -980 +replay_buffer/size 999033 +trainer/num train calls 21000 +trainer/Policy Loss -19.5492 +trainer/Log Pis Mean 25.0291 +trainer/Log Pis Std 13.1858 +trainer/Log Pis Max 76.1924 +trainer/Log Pis Min -4.71348 +trainer/policy/mean Mean -0.032351 +trainer/policy/mean Std 0.901317 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.01916 +trainer/policy/normal/std Std 0.475065 +trainer/policy/normal/std Max 4.6406 +trainer/policy/normal/std Min 0.804863 +trainer/policy/normal/log_std Mean 1.08916 +trainer/policy/normal/log_std Std 0.191997 +trainer/policy/normal/log_std Max 1.53484 +trainer/policy/normal/log_std Min -0.217083 +eval/num steps total 21000 +eval/num paths total 21 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102682 +eval/Actions Std 0.910604 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50983 +time/logging (s) 0.00413098 +time/sampling batch (s) 0.280537 +time/saving (s) 0.00407688 +time/training (s) 6.93088 +time/epoch (s) 9.72945 +time/total (s) 252.445 +Epoch -980 +---------------------------------- --------------- +2022-05-10 13:14:57.352394 PDT | [1] Epoch -979 finished +---------------------------------- --------------- +epoch -979 +replay_buffer/size 999033 +trainer/num train calls 22000 +trainer/Policy Loss -18.9155 +trainer/Log Pis Mean 23.7667 +trainer/Log Pis Std 13.2618 +trainer/Log Pis Max 63.0807 +trainer/Log Pis Min -8.74111 +trainer/policy/mean Mean -0.027716 +trainer/policy/mean Std 0.897038 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.95884 +trainer/policy/normal/std Std 0.464353 +trainer/policy/normal/std Max 4.82194 +trainer/policy/normal/std Min 0.791236 +trainer/policy/normal/log_std Mean 1.06984 +trainer/policy/normal/log_std Std 0.184492 +trainer/policy/normal/log_std Max 1.57318 +trainer/policy/normal/log_std Min -0.23416 +eval/num steps total 22000 +eval/num paths total 22 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116712 +eval/Actions Std 0.911207 +eval/Actions Max 0.99999 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85476 +time/logging (s) 0.00371186 +time/sampling batch (s) 0.530955 +time/saving (s) 0.00355245 +time/training (s) 7.72487 +time/epoch (s) 11.1178 +time/total (s) 263.567 +Epoch -979 +---------------------------------- --------------- +2022-05-10 13:15:07.095272 PDT | [1] Epoch -978 finished +---------------------------------- --------------- +epoch -978 +replay_buffer/size 999033 +trainer/num train calls 23000 +trainer/Policy Loss -18.5718 +trainer/Log Pis Mean 24.3126 +trainer/Log Pis Std 13.032 +trainer/Log Pis Max 73.1459 +trainer/Log Pis Min -10.6713 +trainer/policy/mean Mean -0.0316576 +trainer/policy/mean Std 0.892204 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.03679 +trainer/policy/normal/std Std 0.493245 +trainer/policy/normal/std Max 5.0608 +trainer/policy/normal/std Min 0.837757 +trainer/policy/normal/log_std Mean 1.09498 +trainer/policy/normal/log_std Std 0.188756 +trainer/policy/normal/log_std Max 1.62152 +trainer/policy/normal/log_std Min -0.177028 +eval/num steps total 23000 +eval/num paths total 23 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.341865 +eval/Actions Std 0.865247 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3519 +time/logging (s) 0.00372385 +time/sampling batch (s) 0.781758 +time/saving (s) 0.0034685 +time/training (s) 6.58232 +time/epoch (s) 9.72317 +time/total (s) 273.293 +Epoch -978 +---------------------------------- --------------- +2022-05-10 13:15:16.806565 PDT | [1] Epoch -977 finished +---------------------------------- --------------- +epoch -977 +replay_buffer/size 999033 +trainer/num train calls 24000 +trainer/Policy Loss -19.0907 +trainer/Log Pis Mean 24.2036 +trainer/Log Pis Std 12.9238 +trainer/Log Pis Max 68.8999 +trainer/Log Pis Min -7.69233 +trainer/policy/mean Mean -0.0498419 +trainer/policy/mean Std 0.896758 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 3.01568 +trainer/policy/normal/std Std 0.478031 +trainer/policy/normal/std Max 4.93961 +trainer/policy/normal/std Min 0.726522 +trainer/policy/normal/log_std Mean 1.08823 +trainer/policy/normal/log_std Std 0.189584 +trainer/policy/normal/log_std Max 1.59729 +trainer/policy/normal/log_std Min -0.319487 +eval/num steps total 24000 +eval/num paths total 24 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0926916 +eval/Actions Std 0.900826 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59195 +time/logging (s) 0.00366468 +time/sampling batch (s) 0.277757 +time/saving (s) 0.00343515 +time/training (s) 6.81491 +time/epoch (s) 9.69172 +time/total (s) 282.988 +Epoch -977 +---------------------------------- --------------- +2022-05-10 13:15:26.173626 PDT | [1] Epoch -976 finished +---------------------------------- --------------- +epoch -976 +replay_buffer/size 999033 +trainer/num train calls 25000 +trainer/Policy Loss -18.2468 +trainer/Log Pis Mean 23.9413 +trainer/Log Pis Std 12.6699 +trainer/Log Pis Max 61.6518 +trainer/Log Pis Min -6.59122 +trainer/policy/mean Mean -0.0269313 +trainer/policy/mean Std 0.898236 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.98837 +trainer/policy/normal/std Std 0.480044 +trainer/policy/normal/std Max 5.73097 +trainer/policy/normal/std Min 0.795093 +trainer/policy/normal/log_std Mean 1.07893 +trainer/policy/normal/log_std Std 0.189687 +trainer/policy/normal/log_std Max 1.74588 +trainer/policy/normal/log_std Min -0.229297 +eval/num steps total 25000 +eval/num paths total 25 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0029665 +eval/Actions Std 0.907927 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43046 +time/logging (s) 0.00368719 +time/sampling batch (s) 0.529076 +time/saving (s) 0.00345618 +time/training (s) 6.38084 +time/epoch (s) 9.34752 +time/total (s) 292.338 +Epoch -976 +---------------------------------- --------------- +2022-05-10 13:15:35.893378 PDT | [1] Epoch -975 finished +---------------------------------- --------------- +epoch -975 +replay_buffer/size 999033 +trainer/num train calls 26000 +trainer/Policy Loss -18.8046 +trainer/Log Pis Mean 25.1866 +trainer/Log Pis Std 12.8472 +trainer/Log Pis Max 64.7293 +trainer/Log Pis Min -10.5636 +trainer/policy/mean Mean -0.0350388 +trainer/policy/mean Std 0.900687 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.96404 +trainer/policy/normal/std Std 0.476587 +trainer/policy/normal/std Max 5.12364 +trainer/policy/normal/std Min 0.735826 +trainer/policy/normal/log_std Mean 1.06997 +trainer/policy/normal/log_std Std 0.196828 +trainer/policy/normal/log_std Max 1.63386 +trainer/policy/normal/log_std Min -0.306761 +eval/num steps total 26000 +eval/num paths total 26 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.216369 +eval/Actions Std 0.949657 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43793 +time/logging (s) 0.00408129 +time/sampling batch (s) 0.280872 +time/saving (s) 0.00389938 +time/training (s) 6.97356 +time/epoch (s) 9.70035 +time/total (s) 302.041 +Epoch -975 +---------------------------------- --------------- +2022-05-10 13:15:46.072479 PDT | [1] Epoch -974 finished +---------------------------------- --------------- +epoch -974 +replay_buffer/size 999033 +trainer/num train calls 27000 +trainer/Policy Loss -18.8164 +trainer/Log Pis Mean 23.5712 +trainer/Log Pis Std 12.5124 +trainer/Log Pis Max 60.6599 +trainer/Log Pis Min -6.65577 +trainer/policy/mean Mean -0.0524723 +trainer/policy/mean Std 0.897597 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91004 +trainer/policy/normal/std Std 0.473255 +trainer/policy/normal/std Max 4.75642 +trainer/policy/normal/std Min 0.761439 +trainer/policy/normal/log_std Mean 1.0514 +trainer/policy/normal/log_std Std 0.197155 +trainer/policy/normal/log_std Max 1.5595 +trainer/policy/normal/log_std Min -0.272545 +eval/num steps total 27000 +eval/num paths total 27 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122307 +eval/Actions Std 0.863178 +eval/Actions Max 0.999986 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71832 +time/logging (s) 0.00390695 +time/sampling batch (s) 0.283254 +time/saving (s) 0.00377031 +time/training (s) 7.14942 +time/epoch (s) 10.1587 +time/total (s) 312.203 +Epoch -974 +---------------------------------- --------------- +2022-05-10 13:15:56.232346 PDT | [1] Epoch -973 finished +---------------------------------- --------------- +epoch -973 +replay_buffer/size 999033 +trainer/num train calls 28000 +trainer/Policy Loss -18.692 +trainer/Log Pis Mean 24.3574 +trainer/Log Pis Std 13.0558 +trainer/Log Pis Max 75.3362 +trainer/Log Pis Min -6.37858 +trainer/policy/mean Mean -0.0411185 +trainer/policy/mean Std 0.898262 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.94261 +trainer/policy/normal/std Std 0.479183 +trainer/policy/normal/std Max 6.40153 +trainer/policy/normal/std Min 0.7005 +trainer/policy/normal/log_std Mean 1.06269 +trainer/policy/normal/log_std Std 0.196509 +trainer/policy/normal/log_std Max 1.85654 +trainer/policy/normal/log_std Min -0.355961 +eval/num steps total 28000 +eval/num paths total 28 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.22951 +eval/Actions Std 0.919767 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59693 +time/logging (s) 0.00366634 +time/sampling batch (s) 0.527662 +time/saving (s) 0.00343939 +time/training (s) 7.0083 +time/epoch (s) 10.14 +time/total (s) 322.346 +Epoch -973 +---------------------------------- --------------- +2022-05-10 13:16:06.593776 PDT | [1] Epoch -972 finished +---------------------------------- --------------- +epoch -972 +replay_buffer/size 999033 +trainer/num train calls 29000 +trainer/Policy Loss -19.8647 +trainer/Log Pis Mean 24.4879 +trainer/Log Pis Std 13.0178 +trainer/Log Pis Max 70.5812 +trainer/Log Pis Min -5.56178 +trainer/policy/mean Mean -0.0322994 +trainer/policy/mean Std 0.896901 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.92751 +trainer/policy/normal/std Std 0.46544 +trainer/policy/normal/std Max 5.30809 +trainer/policy/normal/std Min 0.708059 +trainer/policy/normal/log_std Mean 1.0581 +trainer/policy/normal/log_std Std 0.193509 +trainer/policy/normal/log_std Max 1.66923 +trainer/policy/normal/log_std Min -0.345228 +eval/num steps total 29000 +eval/num paths total 29 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00671727 +eval/Actions Std 0.932929 +eval/Actions Max 0.999991 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63181 +time/logging (s) 0.0037328 +time/sampling batch (s) 0.275434 +time/saving (s) 0.00337866 +time/training (s) 7.42804 +time/epoch (s) 10.3424 +time/total (s) 332.692 +Epoch -972 +---------------------------------- --------------- +2022-05-10 13:16:17.182408 PDT | [1] Epoch -971 finished +---------------------------------- --------------- +epoch -971 +replay_buffer/size 999033 +trainer/num train calls 30000 +trainer/Policy Loss -19.7744 +trainer/Log Pis Mean 23.7706 +trainer/Log Pis Std 13.4635 +trainer/Log Pis Max 60.5471 +trainer/Log Pis Min -14.5768 +trainer/policy/mean Mean -0.0239777 +trainer/policy/mean Std 0.898268 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89716 +trainer/policy/normal/std Std 0.475716 +trainer/policy/normal/std Max 4.5765 +trainer/policy/normal/std Min 0.69216 +trainer/policy/normal/log_std Mean 1.04613 +trainer/policy/normal/log_std Std 0.203836 +trainer/policy/normal/log_std Max 1.52094 +trainer/policy/normal/log_std Min -0.367938 +eval/num steps total 30000 +eval/num paths total 30 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0179361 +eval/Actions Std 0.894219 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54772 +time/logging (s) 0.00373372 +time/sampling batch (s) 0.276009 +time/saving (s) 0.00341564 +time/training (s) 7.73847 +time/epoch (s) 10.5693 +time/total (s) 343.264 +Epoch -971 +---------------------------------- --------------- +2022-05-10 13:16:27.618671 PDT | [1] Epoch -970 finished +---------------------------------- --------------- +epoch -970 +replay_buffer/size 999033 +trainer/num train calls 31000 +trainer/Policy Loss -19.4173 +trainer/Log Pis Mean 22.8683 +trainer/Log Pis Std 12.2635 +trainer/Log Pis Max 60.5299 +trainer/Log Pis Min -6.50262 +trainer/policy/mean Mean -0.0295293 +trainer/policy/mean Std 0.902273 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.94813 +trainer/policy/normal/std Std 0.494142 +trainer/policy/normal/std Max 4.66973 +trainer/policy/normal/std Min 0.750808 +trainer/policy/normal/log_std Mean 1.06365 +trainer/policy/normal/log_std Std 0.200759 +trainer/policy/normal/log_std Max 1.5411 +trainer/policy/normal/log_std Min -0.286605 +eval/num steps total 31000 +eval/num paths total 31 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0912263 +eval/Actions Std 0.913732 +eval/Actions Max 0.999984 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5786 +time/logging (s) 0.00371089 +time/sampling batch (s) 0.282146 +time/saving (s) 0.00346729 +time/training (s) 7.54842 +time/epoch (s) 10.4163 +time/total (s) 353.683 +Epoch -970 +---------------------------------- --------------- +2022-05-10 13:16:38.102373 PDT | [1] Epoch -969 finished +---------------------------------- --------------- +epoch -969 +replay_buffer/size 999033 +trainer/num train calls 32000 +trainer/Policy Loss -19.6136 +trainer/Log Pis Mean 23.9338 +trainer/Log Pis Std 13.6392 +trainer/Log Pis Max 67.0373 +trainer/Log Pis Min -8.67184 +trainer/policy/mean Mean -0.0427703 +trainer/policy/mean Std 0.900925 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.93414 +trainer/policy/normal/std Std 0.488053 +trainer/policy/normal/std Max 4.74175 +trainer/policy/normal/std Min 0.653006 +trainer/policy/normal/log_std Mean 1.05874 +trainer/policy/normal/log_std Std 0.203508 +trainer/policy/normal/log_std Max 1.55641 +trainer/policy/normal/log_std Min -0.426169 +eval/num steps total 32000 +eval/num paths total 32 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0916951 +eval/Actions Std 0.874687 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7648 +time/logging (s) 0.00409456 +time/sampling batch (s) 0.278139 +time/saving (s) 0.00395038 +time/training (s) 7.41361 +time/epoch (s) 10.4646 +time/total (s) 364.151 +Epoch -969 +---------------------------------- --------------- +2022-05-10 13:16:48.567481 PDT | [1] Epoch -968 finished +---------------------------------- --------------- +epoch -968 +replay_buffer/size 999033 +trainer/num train calls 33000 +trainer/Policy Loss -18.8126 +trainer/Log Pis Mean 24.271 +trainer/Log Pis Std 13.0087 +trainer/Log Pis Max 63.7674 +trainer/Log Pis Min -10.674 +trainer/policy/mean Mean -0.0334992 +trainer/policy/mean Std 0.894817 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.97441 +trainer/policy/normal/std Std 0.489493 +trainer/policy/normal/std Max 5.69231 +trainer/policy/normal/std Min 0.667563 +trainer/policy/normal/log_std Mean 1.07341 +trainer/policy/normal/log_std Std 0.195063 +trainer/policy/normal/log_std Max 1.73912 +trainer/policy/normal/log_std Min -0.404121 +eval/num steps total 33000 +eval/num paths total 33 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0162731 +eval/Actions Std 0.890446 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68092 +time/logging (s) 0.00376227 +time/sampling batch (s) 0.277843 +time/saving (s) 0.00351815 +time/training (s) 7.47908 +time/epoch (s) 10.4451 +time/total (s) 374.599 +Epoch -968 +---------------------------------- --------------- +2022-05-10 13:16:58.102546 PDT | [1] Epoch -967 finished +---------------------------------- --------------- +epoch -967 +replay_buffer/size 999033 +trainer/num train calls 34000 +trainer/Policy Loss -20.9847 +trainer/Log Pis Mean 23.859 +trainer/Log Pis Std 13.1686 +trainer/Log Pis Max 60.6845 +trainer/Log Pis Min -9.87329 +trainer/policy/mean Mean -0.0395253 +trainer/policy/mean Std 0.89898 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999975 +trainer/policy/normal/std Mean 3.00928 +trainer/policy/normal/std Std 0.510156 +trainer/policy/normal/std Max 5.64338 +trainer/policy/normal/std Min 0.63421 +trainer/policy/normal/log_std Mean 1.08332 +trainer/policy/normal/log_std Std 0.207634 +trainer/policy/normal/log_std Max 1.73048 +trainer/policy/normal/log_std Min -0.455374 +eval/num steps total 34000 +eval/num paths total 34 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00608671 +eval/Actions Std 0.91992 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5752 +time/logging (s) 0.00381308 +time/sampling batch (s) 0.527309 +time/saving (s) 0.00341075 +time/training (s) 6.40608 +time/epoch (s) 9.51581 +time/total (s) 384.117 +Epoch -967 +---------------------------------- --------------- +2022-05-10 13:17:09.427119 PDT | [1] Epoch -966 finished +---------------------------------- --------------- +epoch -966 +replay_buffer/size 999033 +trainer/num train calls 35000 +trainer/Policy Loss -19.3326 +trainer/Log Pis Mean 23.406 +trainer/Log Pis Std 12.5926 +trainer/Log Pis Max 63.3248 +trainer/Log Pis Min -6.11889 +trainer/policy/mean Mean -0.0231684 +trainer/policy/mean Std 0.902129 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89796 +trainer/policy/normal/std Std 0.502309 +trainer/policy/normal/std Max 5.46192 +trainer/policy/normal/std Min 0.703391 +trainer/policy/normal/log_std Mean 1.04476 +trainer/policy/normal/log_std Std 0.212166 +trainer/policy/normal/log_std Max 1.6978 +trainer/policy/normal/log_std Min -0.351843 +eval/num steps total 35000 +eval/num paths total 35 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.144223 +eval/Actions Std 0.90753 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70286 +time/logging (s) 0.00373554 +time/sampling batch (s) 0.282847 +time/saving (s) 0.00346018 +time/training (s) 8.31171 +time/epoch (s) 11.3046 +time/total (s) 395.425 +Epoch -966 +---------------------------------- --------------- +2022-05-10 13:17:18.746461 PDT | [1] Epoch -965 finished +---------------------------------- --------------- +epoch -965 +replay_buffer/size 999033 +trainer/num train calls 36000 +trainer/Policy Loss -18.9019 +trainer/Log Pis Mean 23.9692 +trainer/Log Pis Std 13.5856 +trainer/Log Pis Max 64.8414 +trainer/Log Pis Min -9.8792 +trainer/policy/mean Mean -0.024675 +trainer/policy/mean Std 0.899555 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.9642 +trainer/policy/normal/std Std 0.514564 +trainer/policy/normal/std Max 4.90317 +trainer/policy/normal/std Min 0.680043 +trainer/policy/normal/log_std Mean 1.06723 +trainer/policy/normal/log_std Std 0.213297 +trainer/policy/normal/log_std Max 1.58988 +trainer/policy/normal/log_std Min -0.385599 +eval/num steps total 36000 +eval/num paths total 36 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.194733 +eval/Actions Std 0.942291 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52249 +time/logging (s) 0.00374143 +time/sampling batch (s) 0.280768 +time/saving (s) 0.003465 +time/training (s) 6.48916 +time/epoch (s) 9.29962 +time/total (s) 404.728 +Epoch -965 +---------------------------------- --------------- +2022-05-10 13:17:28.197927 PDT | [1] Epoch -964 finished +---------------------------------- --------------- +epoch -964 +replay_buffer/size 999033 +trainer/num train calls 37000 +trainer/Policy Loss -18.877 +trainer/Log Pis Mean 23.7714 +trainer/Log Pis Std 13.0072 +trainer/Log Pis Max 93.0513 +trainer/Log Pis Min -10.4119 +trainer/policy/mean Mean -0.0313819 +trainer/policy/mean Std 0.898241 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.93559 +trainer/policy/normal/std Std 0.506497 +trainer/policy/normal/std Max 5.74542 +trainer/policy/normal/std Min 0.734384 +trainer/policy/normal/log_std Mean 1.05808 +trainer/policy/normal/log_std Std 0.20943 +trainer/policy/normal/log_std Max 1.7484 +trainer/policy/normal/log_std Min -0.308723 +eval/num steps total 37000 +eval/num paths total 37 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0537044 +eval/Actions Std 0.897155 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75989 +time/logging (s) 0.00409513 +time/sampling batch (s) 0.281328 +time/saving (s) 0.00384551 +time/training (s) 6.38282 +time/epoch (s) 9.43198 +time/total (s) 414.163 +Epoch -964 +---------------------------------- --------------- +2022-05-10 13:17:38.129002 PDT | [1] Epoch -963 finished +---------------------------------- --------------- +epoch -963 +replay_buffer/size 999033 +trainer/num train calls 38000 +trainer/Policy Loss -18.4715 +trainer/Log Pis Mean 23.4892 +trainer/Log Pis Std 13.2656 +trainer/Log Pis Max 68.9192 +trainer/Log Pis Min -10.8375 +trainer/policy/mean Mean -0.0343303 +trainer/policy/mean Std 0.896823 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.92165 +trainer/policy/normal/std Std 0.473907 +trainer/policy/normal/std Max 4.72599 +trainer/policy/normal/std Min 0.681294 +trainer/policy/normal/log_std Mean 1.05522 +trainer/policy/normal/log_std Std 0.199331 +trainer/policy/normal/log_std Max 1.55308 +trainer/policy/normal/log_std Min -0.383762 +eval/num steps total 38000 +eval/num paths total 38 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0364949 +eval/Actions Std 0.882346 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51527 +time/logging (s) 0.0041464 +time/sampling batch (s) 0.280952 +time/saving (s) 0.00373036 +time/training (s) 7.10687 +time/epoch (s) 9.91096 +time/total (s) 424.077 +Epoch -963 +---------------------------------- --------------- +2022-05-10 13:17:48.286309 PDT | [1] Epoch -962 finished +---------------------------------- --------------- +epoch -962 +replay_buffer/size 999033 +trainer/num train calls 39000 +trainer/Policy Loss -19.4496 +trainer/Log Pis Mean 24.3014 +trainer/Log Pis Std 13.6468 +trainer/Log Pis Max 68.6794 +trainer/Log Pis Min -4.18936 +trainer/policy/mean Mean -0.050313 +trainer/policy/mean Std 0.903092 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.90172 +trainer/policy/normal/std Std 0.488022 +trainer/policy/normal/std Max 4.40496 +trainer/policy/normal/std Min 0.59991 +trainer/policy/normal/log_std Mean 1.04677 +trainer/policy/normal/log_std Std 0.209601 +trainer/policy/normal/log_std Max 1.48273 +trainer/policy/normal/log_std Min -0.510976 +eval/num steps total 39000 +eval/num paths total 39 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127757 +eval/Actions Std 0.897136 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69623 +time/logging (s) 0.00394615 +time/sampling batch (s) 0.281745 +time/saving (s) 0.0034456 +time/training (s) 7.15189 +time/epoch (s) 10.1372 +time/total (s) 434.217 +Epoch -962 +---------------------------------- --------------- +2022-05-10 13:17:57.562300 PDT | [1] Epoch -961 finished +---------------------------------- --------------- +epoch -961 +replay_buffer/size 999033 +trainer/num train calls 40000 +trainer/Policy Loss -18.657 +trainer/Log Pis Mean 23.6882 +trainer/Log Pis Std 13.0466 +trainer/Log Pis Max 63.7379 +trainer/Log Pis Min -8.95248 +trainer/policy/mean Mean -0.045949 +trainer/policy/mean Std 0.900389 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8847 +trainer/policy/normal/std Std 0.493651 +trainer/policy/normal/std Max 4.53349 +trainer/policy/normal/std Min 0.650353 +trainer/policy/normal/log_std Mean 1.04008 +trainer/policy/normal/log_std Std 0.214692 +trainer/policy/normal/log_std Max 1.51149 +trainer/policy/normal/log_std Min -0.43024 +eval/num steps total 40000 +eval/num paths total 40 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0912515 +eval/Actions Std 0.905603 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56565 +time/logging (s) 0.0038088 +time/sampling batch (s) 0.277952 +time/saving (s) 0.00340465 +time/training (s) 6.40564 +time/epoch (s) 9.25646 +time/total (s) 443.476 +Epoch -961 +---------------------------------- --------------- +2022-05-10 13:18:07.274372 PDT | [1] Epoch -960 finished +---------------------------------- --------------- +epoch -960 +replay_buffer/size 999033 +trainer/num train calls 41000 +trainer/Policy Loss -19.8332 +trainer/Log Pis Mean 24.6956 +trainer/Log Pis Std 13.5349 +trainer/Log Pis Max 72.6719 +trainer/Log Pis Min -4.6276 +trainer/policy/mean Mean -0.0451671 +trainer/policy/mean Std 0.904441 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.99576 +trainer/policy/normal/std Std 0.50254 +trainer/policy/normal/std Max 5.19247 +trainer/policy/normal/std Min 0.744404 +trainer/policy/normal/log_std Mean 1.07938 +trainer/policy/normal/log_std Std 0.203878 +trainer/policy/normal/log_std Max 1.64721 +trainer/policy/normal/log_std Min -0.295171 +eval/num steps total 41000 +eval/num paths total 41 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.170271 +eval/Actions Std 0.917106 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7802 +time/logging (s) 0.00389093 +time/sampling batch (s) 0.277028 +time/saving (s) 0.00343526 +time/training (s) 6.62802 +time/epoch (s) 9.69258 +time/total (s) 453.172 +Epoch -960 +---------------------------------- --------------- +2022-05-10 13:18:16.936591 PDT | [1] Epoch -959 finished +---------------------------------- --------------- +epoch -959 +replay_buffer/size 999033 +trainer/num train calls 42000 +trainer/Policy Loss -19.3546 +trainer/Log Pis Mean 23.4301 +trainer/Log Pis Std 12.5015 +trainer/Log Pis Max 62.5356 +trainer/Log Pis Min -9.65179 +trainer/policy/mean Mean -0.0351546 +trainer/policy/mean Std 0.901924 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.92432 +trainer/policy/normal/std Std 0.511214 +trainer/policy/normal/std Max 5.96149 +trainer/policy/normal/std Min 0.663777 +trainer/policy/normal/log_std Mean 1.05346 +trainer/policy/normal/log_std Std 0.214674 +trainer/policy/normal/log_std Max 1.78532 +trainer/policy/normal/log_std Min -0.409809 +eval/num steps total 42000 +eval/num paths total 42 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0939041 +eval/Actions Std 0.91383 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64521 +time/logging (s) 0.00370458 +time/sampling batch (s) 0.279552 +time/saving (s) 0.00343432 +time/training (s) 6.71064 +time/epoch (s) 9.64255 +time/total (s) 462.817 +Epoch -959 +---------------------------------- --------------- +2022-05-10 13:18:27.704313 PDT | [1] Epoch -958 finished +---------------------------------- --------------- +epoch -958 +replay_buffer/size 999033 +trainer/num train calls 43000 +trainer/Policy Loss -20.1147 +trainer/Log Pis Mean 25.0837 +trainer/Log Pis Std 13.1563 +trainer/Log Pis Max 63.1721 +trainer/Log Pis Min -5.85326 +trainer/policy/mean Mean -0.0515727 +trainer/policy/mean Std 0.901165 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.91154 +trainer/policy/normal/std Std 0.522512 +trainer/policy/normal/std Max 5.43511 +trainer/policy/normal/std Min 0.649931 +trainer/policy/normal/log_std Mean 1.04738 +trainer/policy/normal/log_std Std 0.225638 +trainer/policy/normal/log_std Max 1.69288 +trainer/policy/normal/log_std Min -0.43089 +eval/num steps total 43000 +eval/num paths total 43 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.482148 +eval/Actions Std 0.792314 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72076 +time/logging (s) 0.00401982 +time/sampling batch (s) 0.526649 +time/saving (s) 0.00394212 +time/training (s) 7.49319 +time/epoch (s) 10.7486 +time/total (s) 473.569 +Epoch -958 +---------------------------------- --------------- +2022-05-10 13:18:37.675759 PDT | [1] Epoch -957 finished +---------------------------------- --------------- +epoch -957 +replay_buffer/size 999033 +trainer/num train calls 44000 +trainer/Policy Loss -19.8223 +trainer/Log Pis Mean 24.2584 +trainer/Log Pis Std 12.381 +trainer/Log Pis Max 64.5854 +trainer/Log Pis Min -5.06821 +trainer/policy/mean Mean -0.0135172 +trainer/policy/mean Std 0.904455 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.88301 +trainer/policy/normal/std Std 0.514721 +trainer/policy/normal/std Max 5.10836 +trainer/policy/normal/std Min 0.619099 +trainer/policy/normal/log_std Mean 1.03854 +trainer/policy/normal/log_std Std 0.21784 +trainer/policy/normal/log_std Max 1.63088 +trainer/policy/normal/log_std Min -0.479491 +eval/num steps total 44000 +eval/num paths total 44 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0996893 +eval/Actions Std 0.8959 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52762 +time/logging (s) 0.00374604 +time/sampling batch (s) 0.775421 +time/saving (s) 0.00346441 +time/training (s) 6.6403 +time/epoch (s) 9.95055 +time/total (s) 483.524 +Epoch -957 +---------------------------------- --------------- +2022-05-10 13:18:48.169304 PDT | [1] Epoch -956 finished +---------------------------------- --------------- +epoch -956 +replay_buffer/size 999033 +trainer/num train calls 45000 +trainer/Policy Loss -18.2944 +trainer/Log Pis Mean 24.1321 +trainer/Log Pis Std 13.2899 +trainer/Log Pis Max 69.089 +trainer/Log Pis Min -9.59884 +trainer/policy/mean Mean -0.0365966 +trainer/policy/mean Std 0.900275 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.94234 +trainer/policy/normal/std Std 0.506147 +trainer/policy/normal/std Max 5.17017 +trainer/policy/normal/std Min 0.618638 +trainer/policy/normal/log_std Mean 1.06021 +trainer/policy/normal/log_std Std 0.21103 +trainer/policy/normal/log_std Max 1.64291 +trainer/policy/normal/log_std Min -0.480234 +eval/num steps total 45000 +eval/num paths total 45 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00295422 +eval/Actions Std 0.962332 +eval/Actions Max 0.999997 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4611 +time/logging (s) 0.00389829 +time/sampling batch (s) 0.274891 +time/saving (s) 0.00337559 +time/training (s) 7.73101 +time/epoch (s) 10.4743 +time/total (s) 494.001 +Epoch -956 +---------------------------------- --------------- +2022-05-10 13:18:58.656505 PDT | [1] Epoch -955 finished +---------------------------------- --------------- +epoch -955 +replay_buffer/size 999033 +trainer/num train calls 46000 +trainer/Policy Loss -18.7156 +trainer/Log Pis Mean 25.1636 +trainer/Log Pis Std 13.0707 +trainer/Log Pis Max 61.3013 +trainer/Log Pis Min -5.44886 +trainer/policy/mean Mean -0.0448691 +trainer/policy/mean Std 0.904324 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.94043 +trainer/policy/normal/std Std 0.52721 +trainer/policy/normal/std Max 5.122 +trainer/policy/normal/std Min 0.726998 +trainer/policy/normal/log_std Mean 1.05789 +trainer/policy/normal/log_std Std 0.220198 +trainer/policy/normal/log_std Max 1.63355 +trainer/policy/normal/log_std Min -0.318831 +eval/num steps total 46000 +eval/num paths total 46 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0294859 +eval/Actions Std 0.916234 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6475 +time/logging (s) 0.00398028 +time/sampling batch (s) 0.533706 +time/saving (s) 0.00532565 +time/training (s) 7.27709 +time/epoch (s) 10.4676 +time/total (s) 504.471 +Epoch -955 +---------------------------------- --------------- +2022-05-10 13:19:09.215014 PDT | [1] Epoch -954 finished +---------------------------------- --------------- +epoch -954 +replay_buffer/size 999033 +trainer/num train calls 47000 +trainer/Policy Loss -20.6862 +trainer/Log Pis Mean 25.0723 +trainer/Log Pis Std 12.914 +trainer/Log Pis Max 71.2178 +trainer/Log Pis Min -9.69294 +trainer/policy/mean Mean -0.0408756 +trainer/policy/mean Std 0.903311 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89722 +trainer/policy/normal/std Std 0.501816 +trainer/policy/normal/std Max 4.87709 +trainer/policy/normal/std Min 0.589641 +trainer/policy/normal/log_std Mean 1.0442 +trainer/policy/normal/log_std Std 0.215472 +trainer/policy/normal/log_std Max 1.58455 +trainer/policy/normal/log_std Min -0.528241 +eval/num steps total 47000 +eval/num paths total 47 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.408962 +eval/Actions Std 0.830814 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80126 +time/logging (s) 0.00372973 +time/sampling batch (s) 0.527703 +time/saving (s) 0.00347293 +time/training (s) 7.20259 +time/epoch (s) 10.5388 +time/total (s) 515.013 +Epoch -954 +---------------------------------- --------------- +2022-05-10 13:19:20.690270 PDT | [1] Epoch -953 finished +---------------------------------- --------------- +epoch -953 +replay_buffer/size 999033 +trainer/num train calls 48000 +trainer/Policy Loss -18.3097 +trainer/Log Pis Mean 25.0174 +trainer/Log Pis Std 14.3468 +trainer/Log Pis Max 71.0505 +trainer/Log Pis Min -8.21566 +trainer/policy/mean Mean -0.0266665 +trainer/policy/mean Std 0.899818 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.94143 +trainer/policy/normal/std Std 0.501444 +trainer/policy/normal/std Max 4.78645 +trainer/policy/normal/std Min 0.679704 +trainer/policy/normal/log_std Mean 1.06029 +trainer/policy/normal/log_std Std 0.208605 +trainer/policy/normal/log_std Max 1.56579 +trainer/policy/normal/log_std Min -0.386097 +eval/num steps total 48000 +eval/num paths total 48 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0597002 +eval/Actions Std 0.898759 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70076 +time/logging (s) 0.0042905 +time/sampling batch (s) 0.529625 +time/saving (s) 0.00393902 +time/training (s) 8.21749 +time/epoch (s) 11.4561 +time/total (s) 526.472 +Epoch -953 +---------------------------------- --------------- +2022-05-10 13:19:30.350576 PDT | [1] Epoch -952 finished +---------------------------------- --------------- +epoch -952 +replay_buffer/size 999033 +trainer/num train calls 49000 +trainer/Policy Loss -18.9619 +trainer/Log Pis Mean 24.8179 +trainer/Log Pis Std 13.214 +trainer/Log Pis Max 72.5912 +trainer/Log Pis Min -4.93072 +trainer/policy/mean Mean -0.0542067 +trainer/policy/mean Std 0.902282 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.90629 +trainer/policy/normal/std Std 0.527224 +trainer/policy/normal/std Max 5.09287 +trainer/policy/normal/std Min 0.574061 +trainer/policy/normal/log_std Mean 1.04538 +trainer/policy/normal/log_std Std 0.226104 +trainer/policy/normal/log_std Max 1.62784 +trainer/policy/normal/log_std Min -0.555019 +eval/num steps total 49000 +eval/num paths total 49 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.193111 +eval/Actions Std 0.952796 +eval/Actions Max 0.999991 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48065 +time/logging (s) 0.00379756 +time/sampling batch (s) 0.278754 +time/saving (s) 0.00359735 +time/training (s) 6.87294 +time/epoch (s) 9.63974 +time/total (s) 536.115 +Epoch -952 +---------------------------------- --------------- +2022-05-10 13:19:40.016888 PDT | [1] Epoch -951 finished +---------------------------------- --------------- +epoch -951 +replay_buffer/size 999033 +trainer/num train calls 50000 +trainer/Policy Loss -19.0532 +trainer/Log Pis Mean 24.5281 +trainer/Log Pis Std 13.1941 +trainer/Log Pis Max 83.3499 +trainer/Log Pis Min -7.92936 +trainer/policy/mean Mean -0.0482541 +trainer/policy/mean Std 0.899093 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8948 +trainer/policy/normal/std Std 0.511356 +trainer/policy/normal/std Max 4.82787 +trainer/policy/normal/std Min 0.646486 +trainer/policy/normal/log_std Mean 1.04212 +trainer/policy/normal/log_std Std 0.223132 +trainer/policy/normal/log_std Max 1.57441 +trainer/policy/normal/log_std Min -0.436203 +eval/num steps total 50000 +eval/num paths total 50 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0725885 +eval/Actions Std 0.904012 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7838 +time/logging (s) 0.00375939 +time/sampling batch (s) 0.278206 +time/saving (s) 0.00337777 +time/training (s) 6.57755 +time/epoch (s) 9.64669 +time/total (s) 545.765 +Epoch -951 +---------------------------------- --------------- +2022-05-10 13:19:51.193399 PDT | [1] Epoch -950 finished +---------------------------------- --------------- +epoch -950 +replay_buffer/size 999033 +trainer/num train calls 51000 +trainer/Policy Loss -18.9842 +trainer/Log Pis Mean 24.7049 +trainer/Log Pis Std 13.3717 +trainer/Log Pis Max 66.5506 +trainer/Log Pis Min -6.07317 +trainer/policy/mean Mean -0.0411357 +trainer/policy/mean Std 0.90429 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.94037 +trainer/policy/normal/std Std 0.509641 +trainer/policy/normal/std Max 5.26027 +trainer/policy/normal/std Min 0.644556 +trainer/policy/normal/log_std Mean 1.0593 +trainer/policy/normal/log_std Std 0.212296 +trainer/policy/normal/log_std Max 1.66018 +trainer/policy/normal/log_std Min -0.439193 +eval/num steps total 51000 +eval/num paths total 51 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.284024 +eval/Actions Std 0.895531 +eval/Actions Max 0.999982 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81329 +time/logging (s) 0.00365867 +time/sampling batch (s) 0.279902 +time/saving (s) 0.00338433 +time/training (s) 8.05646 +time/epoch (s) 11.1567 +time/total (s) 556.924 +Epoch -950 +---------------------------------- --------------- +2022-05-10 13:20:01.122479 PDT | [1] Epoch -949 finished +---------------------------------- --------------- +epoch -949 +replay_buffer/size 999033 +trainer/num train calls 52000 +trainer/Policy Loss -19.4907 +trainer/Log Pis Mean 23.9698 +trainer/Log Pis Std 12.2975 +trainer/Log Pis Max 66.6483 +trainer/Log Pis Min -6.7973 +trainer/policy/mean Mean -0.0385061 +trainer/policy/mean Std 0.899263 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.93481 +trainer/policy/normal/std Std 0.514608 +trainer/policy/normal/std Max 5.25416 +trainer/policy/normal/std Min 0.627244 +trainer/policy/normal/log_std Mean 1.05614 +trainer/policy/normal/log_std Std 0.221753 +trainer/policy/normal/log_std Max 1.65902 +trainer/policy/normal/log_std Min -0.46642 +eval/num steps total 52000 +eval/num paths total 52 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0465845 +eval/Actions Std 0.906816 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71528 +time/logging (s) 0.00382969 +time/sampling batch (s) 0.291178 +time/saving (s) 0.00343962 +time/training (s) 6.8954 +time/epoch (s) 9.90913 +time/total (s) 566.836 +Epoch -949 +---------------------------------- --------------- +2022-05-10 13:20:11.307443 PDT | [1] Epoch -948 finished +---------------------------------- --------------- +epoch -948 +replay_buffer/size 999033 +trainer/num train calls 53000 +trainer/Policy Loss -19.6114 +trainer/Log Pis Mean 24.3432 +trainer/Log Pis Std 13.0996 +trainer/Log Pis Max 64.7624 +trainer/Log Pis Min -9.17654 +trainer/policy/mean Mean -0.0277857 +trainer/policy/mean Std 0.906745 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8981 +trainer/policy/normal/std Std 0.51408 +trainer/policy/normal/std Max 4.79633 +trainer/policy/normal/std Min 0.653008 +trainer/policy/normal/log_std Mean 1.04367 +trainer/policy/normal/log_std Std 0.218957 +trainer/policy/normal/log_std Max 1.56785 +trainer/policy/normal/log_std Min -0.426165 +eval/num steps total 53000 +eval/num paths total 53 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.015754 +eval/Actions Std 0.907808 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64689 +time/logging (s) 0.00410305 +time/sampling batch (s) 0.279666 +time/saving (s) 0.00382673 +time/training (s) 7.2309 +time/epoch (s) 10.1654 +time/total (s) 577.005 +Epoch -948 +---------------------------------- --------------- +2022-05-10 13:20:21.295216 PDT | [1] Epoch -947 finished +---------------------------------- --------------- +epoch -947 +replay_buffer/size 999033 +trainer/num train calls 54000 +trainer/Policy Loss -18.9598 +trainer/Log Pis Mean 24.265 +trainer/Log Pis Std 13.0957 +trainer/Log Pis Max 64.3997 +trainer/Log Pis Min -8.47533 +trainer/policy/mean Mean -0.0401227 +trainer/policy/mean Std 0.89942 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.92297 +trainer/policy/normal/std Std 0.522927 +trainer/policy/normal/std Max 6.88868 +trainer/policy/normal/std Min 0.627973 +trainer/policy/normal/log_std Mean 1.05177 +trainer/policy/normal/log_std Std 0.222477 +trainer/policy/normal/log_std Max 1.92988 +trainer/policy/normal/log_std Min -0.465258 +eval/num steps total 54000 +eval/num paths total 54 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.314095 +eval/Actions Std 0.86986 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.95042 +time/logging (s) 0.00391192 +time/sampling batch (s) 0.279219 +time/saving (s) 0.00367201 +time/training (s) 6.73041 +time/epoch (s) 9.96763 +time/total (s) 586.976 +Epoch -947 +---------------------------------- --------------- +2022-05-10 13:20:31.518402 PDT | [1] Epoch -946 finished +---------------------------------- --------------- +epoch -946 +replay_buffer/size 999033 +trainer/num train calls 55000 +trainer/Policy Loss -19.8792 +trainer/Log Pis Mean 25.1454 +trainer/Log Pis Std 12.7099 +trainer/Log Pis Max 80.4601 +trainer/Log Pis Min -4.84019 +trainer/policy/mean Mean -0.0372234 +trainer/policy/mean Std 0.90282 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.94252 +trainer/policy/normal/std Std 0.534663 +trainer/policy/normal/std Max 4.75441 +trainer/policy/normal/std Min 0.591263 +trainer/policy/normal/log_std Mean 1.05736 +trainer/policy/normal/log_std Std 0.229123 +trainer/policy/normal/log_std Max 1.55907 +trainer/policy/normal/log_std Min -0.525495 +eval/num steps total 55000 +eval/num paths total 55 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116985 +eval/Actions Std 0.891926 +eval/Actions Max 0.999981 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73224 +time/logging (s) 0.00385415 +time/sampling batch (s) 0.532853 +time/saving (s) 0.0034619 +time/training (s) 6.93071 +time/epoch (s) 10.2031 +time/total (s) 597.182 +Epoch -946 +---------------------------------- --------------- +2022-05-10 13:20:41.471404 PDT | [1] Epoch -945 finished +---------------------------------- --------------- +epoch -945 +replay_buffer/size 999033 +trainer/num train calls 56000 +trainer/Policy Loss -19.0931 +trainer/Log Pis Mean 24.5287 +trainer/Log Pis Std 12.9226 +trainer/Log Pis Max 66.2809 +trainer/Log Pis Min -9.30626 +trainer/policy/mean Mean -0.0116563 +trainer/policy/mean Std 0.897949 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.92564 +trainer/policy/normal/std Std 0.523957 +trainer/policy/normal/std Max 4.74604 +trainer/policy/normal/std Min 0.593467 +trainer/policy/normal/log_std Mean 1.05212 +trainer/policy/normal/log_std Std 0.226651 +trainer/policy/normal/log_std Max 1.55731 +trainer/policy/normal/log_std Min -0.521774 +eval/num steps total 56000 +eval/num paths total 56 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0716442 +eval/Actions Std 0.91589 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6495 +time/logging (s) 0.0038259 +time/sampling batch (s) 0.283534 +time/saving (s) 0.00344234 +time/training (s) 6.99285 +time/epoch (s) 9.93315 +time/total (s) 607.118 +Epoch -945 +---------------------------------- --------------- +2022-05-10 13:20:52.501364 PDT | [1] Epoch -944 finished +---------------------------------- --------------- +epoch -944 +replay_buffer/size 999033 +trainer/num train calls 57000 +trainer/Policy Loss -20.456 +trainer/Log Pis Mean 24.6186 +trainer/Log Pis Std 13.2672 +trainer/Log Pis Max 78.6536 +trainer/Log Pis Min -8.73427 +trainer/policy/mean Mean -0.0216464 +trainer/policy/mean Std 0.905783 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.95656 +trainer/policy/normal/std Std 0.55463 +trainer/policy/normal/std Max 4.92011 +trainer/policy/normal/std Min 0.611464 +trainer/policy/normal/log_std Mean 1.06033 +trainer/policy/normal/log_std Std 0.238807 +trainer/policy/normal/log_std Max 1.59333 +trainer/policy/normal/log_std Min -0.491899 +eval/num steps total 57000 +eval/num paths total 57 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122403 +eval/Actions Std 0.899596 +eval/Actions Max 0.999992 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86503 +time/logging (s) 0.00368021 +time/sampling batch (s) 0.534746 +time/saving (s) 0.0033785 +time/training (s) 7.60282 +time/epoch (s) 11.0097 +time/total (s) 618.13 +Epoch -944 +---------------------------------- --------------- +2022-05-10 13:21:02.637024 PDT | [1] Epoch -943 finished +---------------------------------- --------------- +epoch -943 +replay_buffer/size 999033 +trainer/num train calls 58000 +trainer/Policy Loss -19.7227 +trainer/Log Pis Mean 24.4653 +trainer/Log Pis Std 12.6565 +trainer/Log Pis Max 57.2041 +trainer/Log Pis Min -3.13312 +trainer/policy/mean Mean -0.0448262 +trainer/policy/mean Std 0.901698 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.97077 +trainer/policy/normal/std Std 0.559676 +trainer/policy/normal/std Max 5.47951 +trainer/policy/normal/std Min 0.559369 +trainer/policy/normal/log_std Mean 1.06523 +trainer/policy/normal/log_std Std 0.238267 +trainer/policy/normal/log_std Max 1.70101 +trainer/policy/normal/log_std Min -0.580945 +eval/num steps total 58000 +eval/num paths total 58 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0191581 +eval/Actions Std 0.896797 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62747 +time/logging (s) 0.00382382 +time/sampling batch (s) 0.283541 +time/saving (s) 0.00340094 +time/training (s) 7.19771 +time/epoch (s) 10.1159 +time/total (s) 628.249 +Epoch -943 +---------------------------------- --------------- +2022-05-10 13:21:12.362940 PDT | [1] Epoch -942 finished +---------------------------------- --------------- +epoch -942 +replay_buffer/size 999033 +trainer/num train calls 59000 +trainer/Policy Loss -19.3921 +trainer/Log Pis Mean 24.2559 +trainer/Log Pis Std 12.5237 +trainer/Log Pis Max 64.8346 +trainer/Log Pis Min -6.89775 +trainer/policy/mean Mean -0.0476234 +trainer/policy/mean Std 0.898975 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.91486 +trainer/policy/normal/std Std 0.539548 +trainer/policy/normal/std Max 6.07327 +trainer/policy/normal/std Min 0.498046 +trainer/policy/normal/log_std Mean 1.0481 +trainer/policy/normal/log_std Std 0.225813 +trainer/policy/normal/log_std Max 1.8039 +trainer/policy/normal/log_std Min -0.697062 +eval/num steps total 59000 +eval/num paths total 59 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.359314 +eval/Actions Std 0.862806 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67368 +time/logging (s) 0.00393732 +time/sampling batch (s) 0.532174 +time/saving (s) 0.00382578 +time/training (s) 6.49247 +time/epoch (s) 9.70608 +time/total (s) 637.958 +Epoch -942 +---------------------------------- --------------- +2022-05-10 13:21:23.367488 PDT | [1] Epoch -941 finished +---------------------------------- --------------- +epoch -941 +replay_buffer/size 999033 +trainer/num train calls 60000 +trainer/Policy Loss -18.5103 +trainer/Log Pis Mean 24.044 +trainer/Log Pis Std 12.9041 +trainer/Log Pis Max 62.5257 +trainer/Log Pis Min -8.72979 +trainer/policy/mean Mean -0.0186593 +trainer/policy/mean Std 0.901937 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.96228 +trainer/policy/normal/std Std 0.531432 +trainer/policy/normal/std Max 5.1488 +trainer/policy/normal/std Min 0.574275 +trainer/policy/normal/log_std Mean 1.06446 +trainer/policy/normal/log_std Std 0.227884 +trainer/policy/normal/log_std Max 1.63876 +trainer/policy/normal/log_std Min -0.554648 +eval/num steps total 60000 +eval/num paths total 60 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0153764 +eval/Actions Std 0.889843 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62887 +time/logging (s) 0.00370794 +time/sampling batch (s) 0.281783 +time/saving (s) 0.00350136 +time/training (s) 8.06628 +time/epoch (s) 10.9841 +time/total (s) 648.945 +Epoch -941 +---------------------------------- --------------- +2022-05-10 13:21:33.248065 PDT | [1] Epoch -940 finished +---------------------------------- --------------- +epoch -940 +replay_buffer/size 999033 +trainer/num train calls 61000 +trainer/Policy Loss -19.1815 +trainer/Log Pis Mean 26.194 +trainer/Log Pis Std 13.5853 +trainer/Log Pis Max 68.658 +trainer/Log Pis Min -5.91661 +trainer/policy/mean Mean -0.0538153 +trainer/policy/mean Std 0.90436 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.92382 +trainer/policy/normal/std Std 0.530382 +trainer/policy/normal/std Max 5.23987 +trainer/policy/normal/std Min 0.577708 +trainer/policy/normal/log_std Mean 1.0514 +trainer/policy/normal/log_std Std 0.226125 +trainer/policy/normal/log_std Max 1.6563 +trainer/policy/normal/log_std Min -0.548688 +eval/num steps total 61000 +eval/num paths total 61 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125887 +eval/Actions Std 0.911475 +eval/Actions Max 0.999978 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61858 +time/logging (s) 0.00365303 +time/sampling batch (s) 0.533079 +time/saving (s) 0.00342649 +time/training (s) 6.70175 +time/epoch (s) 9.86049 +time/total (s) 658.809 +Epoch -940 +---------------------------------- --------------- +2022-05-10 13:21:42.921391 PDT | [1] Epoch -939 finished +---------------------------------- --------------- +epoch -939 +replay_buffer/size 999033 +trainer/num train calls 62000 +trainer/Policy Loss -18.5675 +trainer/Log Pis Mean 24.6069 +trainer/Log Pis Std 13.0209 +trainer/Log Pis Max 70.2164 +trainer/Log Pis Min -5.96034 +trainer/policy/mean Mean -0.0320046 +trainer/policy/mean Std 0.901396 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.94701 +trainer/policy/normal/std Std 0.551182 +trainer/policy/normal/std Max 6.04064 +trainer/policy/normal/std Min 0.566545 +trainer/policy/normal/log_std Mean 1.05859 +trainer/policy/normal/log_std Std 0.228241 +trainer/policy/normal/log_std Max 1.79851 +trainer/policy/normal/log_std Min -0.568199 +eval/num steps total 62000 +eval/num paths total 62 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127291 +eval/Actions Std 0.891701 +eval/Actions Max 0.999981 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70994 +time/logging (s) 0.00365844 +time/sampling batch (s) 0.279036 +time/saving (s) 0.00343412 +time/training (s) 6.65745 +time/epoch (s) 9.65352 +time/total (s) 668.465 +Epoch -939 +---------------------------------- --------------- +2022-05-10 13:21:54.141558 PDT | [1] Epoch -938 finished +---------------------------------- --------------- +epoch -938 +replay_buffer/size 999033 +trainer/num train calls 63000 +trainer/Policy Loss -20.0531 +trainer/Log Pis Mean 24.5545 +trainer/Log Pis Std 13.2643 +trainer/Log Pis Max 67.1782 +trainer/Log Pis Min -4.77679 +trainer/policy/mean Mean -0.041898 +trainer/policy/mean Std 0.900585 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87978 +trainer/policy/normal/std Std 0.532061 +trainer/policy/normal/std Max 5.05077 +trainer/policy/normal/std Min 0.536517 +trainer/policy/normal/log_std Mean 1.03514 +trainer/policy/normal/log_std Std 0.232375 +trainer/policy/normal/log_std Max 1.61954 +trainer/policy/normal/log_std Min -0.622657 +eval/num steps total 63000 +eval/num paths total 63 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0751675 +eval/Actions Std 0.908482 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70131 +time/logging (s) 0.00376143 +time/sampling batch (s) 0.282737 +time/saving (s) 0.00344086 +time/training (s) 8.20904 +time/epoch (s) 11.2003 +time/total (s) 679.669 +Epoch -938 +---------------------------------- --------------- +2022-05-10 13:22:03.303360 PDT | [1] Epoch -937 finished +---------------------------------- --------------- +epoch -937 +replay_buffer/size 999033 +trainer/num train calls 64000 +trainer/Policy Loss -20.1292 +trainer/Log Pis Mean 23.5194 +trainer/Log Pis Std 12.6275 +trainer/Log Pis Max 68.6881 +trainer/Log Pis Min -9.99444 +trainer/policy/mean Mean -0.0468384 +trainer/policy/mean Std 0.901421 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85631 +trainer/policy/normal/std Std 0.522684 +trainer/policy/normal/std Max 4.9793 +trainer/policy/normal/std Min 0.574108 +trainer/policy/normal/log_std Mean 1.02726 +trainer/policy/normal/log_std Std 0.231277 +trainer/policy/normal/log_std Max 1.60529 +trainer/policy/normal/log_std Min -0.554938 +eval/num steps total 64000 +eval/num paths total 64 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103903 +eval/Actions Std 0.892859 +eval/Actions Max 0.999985 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6508 +time/logging (s) 0.00412672 +time/sampling batch (s) 0.278561 +time/saving (s) 0.00378347 +time/training (s) 6.2051 +time/epoch (s) 9.14238 +time/total (s) 688.814 +Epoch -937 +---------------------------------- --------------- +2022-05-10 13:22:14.527933 PDT | [1] Epoch -936 finished +---------------------------------- --------------- +epoch -936 +replay_buffer/size 999033 +trainer/num train calls 65000 +trainer/Policy Loss -18.6276 +trainer/Log Pis Mean 24.0132 +trainer/Log Pis Std 13.7845 +trainer/Log Pis Max 76.2168 +trainer/Log Pis Min -10.675 +trainer/policy/mean Mean -0.03229 +trainer/policy/mean Std 0.90388 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88151 +trainer/policy/normal/std Std 0.51715 +trainer/policy/normal/std Max 4.98507 +trainer/policy/normal/std Min 0.562281 +trainer/policy/normal/log_std Mean 1.03685 +trainer/policy/normal/log_std Std 0.227279 +trainer/policy/normal/log_std Max 1.60645 +trainer/policy/normal/log_std Min -0.575753 +eval/num steps total 65000 +eval/num paths total 65 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.341844 +eval/Actions Std 0.8709 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86699 +time/logging (s) 0.00380035 +time/sampling batch (s) 0.527975 +time/saving (s) 0.00367308 +time/training (s) 7.80183 +time/epoch (s) 11.2043 +time/total (s) 700.021 +Epoch -936 +---------------------------------- --------------- +2022-05-10 13:22:25.402235 PDT | [1] Epoch -935 finished +---------------------------------- --------------- +epoch -935 +replay_buffer/size 999033 +trainer/num train calls 66000 +trainer/Policy Loss -19.6154 +trainer/Log Pis Mean 24.6467 +trainer/Log Pis Std 13.6092 +trainer/Log Pis Max 73.4745 +trainer/Log Pis Min -11.6124 +trainer/policy/mean Mean -0.0318727 +trainer/policy/mean Std 0.901955 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.92586 +trainer/policy/normal/std Std 0.53347 +trainer/policy/normal/std Max 5.38265 +trainer/policy/normal/std Min 0.549007 +trainer/policy/normal/log_std Mean 1.05162 +trainer/policy/normal/log_std Std 0.229402 +trainer/policy/normal/log_std Max 1.68318 +trainer/policy/normal/log_std Min -0.599643 +eval/num steps total 66000 +eval/num paths total 66 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0232042 +eval/Actions Std 0.939709 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65421 +time/logging (s) 0.00369573 +time/sampling batch (s) 0.527723 +time/saving (s) 0.00341992 +time/training (s) 7.66548 +time/epoch (s) 10.8545 +time/total (s) 710.879 +Epoch -935 +---------------------------------- --------------- +2022-05-10 13:22:35.555278 PDT | [1] Epoch -934 finished +---------------------------------- --------------- +epoch -934 +replay_buffer/size 999033 +trainer/num train calls 67000 +trainer/Policy Loss -19.57 +trainer/Log Pis Mean 23.8406 +trainer/Log Pis Std 13.4526 +trainer/Log Pis Max 75.125 +trainer/Log Pis Min -8.0824 +trainer/policy/mean Mean -0.0321258 +trainer/policy/mean Std 0.90117 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8783 +trainer/policy/normal/std Std 0.532156 +trainer/policy/normal/std Max 5.17401 +trainer/policy/normal/std Min 0.527969 +trainer/policy/normal/log_std Mean 1.03369 +trainer/policy/normal/log_std Std 0.239919 +trainer/policy/normal/log_std Max 1.64365 +trainer/policy/normal/log_std Min -0.638718 +eval/num steps total 67000 +eval/num paths total 67 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10541 +eval/Actions Std 0.908215 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61668 +time/logging (s) 0.00379638 +time/sampling batch (s) 0.275473 +time/saving (s) 0.00342263 +time/training (s) 7.23405 +time/epoch (s) 10.1334 +time/total (s) 721.015 +Epoch -934 +---------------------------------- --------------- +2022-05-10 13:22:46.138151 PDT | [1] Epoch -933 finished +---------------------------------- --------------- +epoch -933 +replay_buffer/size 999033 +trainer/num train calls 68000 +trainer/Policy Loss -20.3233 +trainer/Log Pis Mean 24.2079 +trainer/Log Pis Std 12.7526 +trainer/Log Pis Max 69.3437 +trainer/Log Pis Min -4.24754 +trainer/policy/mean Mean -0.0304783 +trainer/policy/mean Std 0.902261 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.93834 +trainer/policy/normal/std Std 0.528581 +trainer/policy/normal/std Max 5.10184 +trainer/policy/normal/std Min 0.53503 +trainer/policy/normal/log_std Mean 1.05635 +trainer/policy/normal/log_std Std 0.227001 +trainer/policy/normal/log_std Max 1.6296 +trainer/policy/normal/log_std Min -0.625432 +eval/num steps total 68000 +eval/num paths total 68 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0946129 +eval/Actions Std 0.909998 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58435 +time/logging (s) 0.00368172 +time/sampling batch (s) 0.526244 +time/saving (s) 0.00345112 +time/training (s) 7.4455 +time/epoch (s) 10.5632 +time/total (s) 731.582 +Epoch -933 +---------------------------------- --------------- +2022-05-10 13:22:56.998028 PDT | [1] Epoch -932 finished +---------------------------------- --------------- +epoch -932 +replay_buffer/size 999033 +trainer/num train calls 69000 +trainer/Policy Loss -19.0052 +trainer/Log Pis Mean 23.6171 +trainer/Log Pis Std 12.4669 +trainer/Log Pis Max 60.111 +trainer/Log Pis Min -4.56622 +trainer/policy/mean Mean -0.0498964 +trainer/policy/mean Std 0.902362 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.87276 +trainer/policy/normal/std Std 0.535039 +trainer/policy/normal/std Max 5.28847 +trainer/policy/normal/std Min 0.529225 +trainer/policy/normal/log_std Mean 1.03201 +trainer/policy/normal/log_std Std 0.237453 +trainer/policy/normal/log_std Max 1.66553 +trainer/policy/normal/log_std Min -0.636342 +eval/num steps total 69000 +eval/num paths total 69 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0853422 +eval/Actions Std 0.873893 +eval/Actions Max 0.999989 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63476 +time/logging (s) 0.00406356 +time/sampling batch (s) 0.526505 +time/saving (s) 0.00376907 +time/training (s) 7.67162 +time/epoch (s) 10.8407 +time/total (s) 742.425 +Epoch -932 +---------------------------------- --------------- +2022-05-10 13:23:07.725513 PDT | [1] Epoch -931 finished +---------------------------------- --------------- +epoch -931 +replay_buffer/size 999033 +trainer/num train calls 70000 +trainer/Policy Loss -19.1718 +trainer/Log Pis Mean 24.7402 +trainer/Log Pis Std 12.7343 +trainer/Log Pis Max 79.6911 +trainer/Log Pis Min -3.37205 +trainer/policy/mean Mean -0.0450629 +trainer/policy/mean Std 0.903014 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89039 +trainer/policy/normal/std Std 0.523369 +trainer/policy/normal/std Max 6.05547 +trainer/policy/normal/std Min 0.547413 +trainer/policy/normal/log_std Mean 1.03982 +trainer/policy/normal/log_std Std 0.227276 +trainer/policy/normal/log_std Max 1.80096 +trainer/policy/normal/log_std Min -0.602551 +eval/num steps total 70000 +eval/num paths total 70 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11635 +eval/Actions Std 0.904506 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59264 +time/logging (s) 0.00382854 +time/sampling batch (s) 0.279456 +time/saving (s) 0.00372753 +time/training (s) 7.82739 +time/epoch (s) 10.707 +time/total (s) 753.135 +Epoch -931 +---------------------------------- --------------- +2022-05-10 13:23:18.362640 PDT | [1] Epoch -930 finished +---------------------------------- --------------- +epoch -930 +replay_buffer/size 999033 +trainer/num train calls 71000 +trainer/Policy Loss -18.9851 +trainer/Log Pis Mean 24.2484 +trainer/Log Pis Std 13.3091 +trainer/Log Pis Max 69.5487 +trainer/Log Pis Min -7.49755 +trainer/policy/mean Mean -0.015187 +trainer/policy/mean Std 0.90696 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87888 +trainer/policy/normal/std Std 0.541937 +trainer/policy/normal/std Max 6.07828 +trainer/policy/normal/std Min 0.540189 +trainer/policy/normal/log_std Mean 1.03406 +trainer/policy/normal/log_std Std 0.236054 +trainer/policy/normal/log_std Max 1.80472 +trainer/policy/normal/log_std Min -0.615837 +eval/num steps total 71000 +eval/num paths total 71 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.018758 +eval/Actions Std 0.889882 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51023 +time/logging (s) 0.00390931 +time/sampling batch (s) 0.282838 +time/saving (s) 0.00347124 +time/training (s) 7.81663 +time/epoch (s) 10.6171 +time/total (s) 763.755 +Epoch -930 +---------------------------------- --------------- +2022-05-10 13:23:28.980749 PDT | [1] Epoch -929 finished +---------------------------------- --------------- +epoch -929 +replay_buffer/size 999033 +trainer/num train calls 72000 +trainer/Policy Loss -19.8817 +trainer/Log Pis Mean 23.9129 +trainer/Log Pis Std 13.1886 +trainer/Log Pis Max 66.4354 +trainer/Log Pis Min -7.87155 +trainer/policy/mean Mean -0.0523275 +trainer/policy/mean Std 0.905409 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85066 +trainer/policy/normal/std Std 0.52597 +trainer/policy/normal/std Max 4.62914 +trainer/policy/normal/std Min 0.574186 +trainer/policy/normal/log_std Mean 1.0243 +trainer/policy/normal/log_std Std 0.237654 +trainer/policy/normal/log_std Max 1.53237 +trainer/policy/normal/log_std Min -0.554802 +eval/num steps total 72000 +eval/num paths total 72 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0788594 +eval/Actions Std 0.881556 +eval/Actions Max 0.999984 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65887 +time/logging (s) 0.00367907 +time/sampling batch (s) 0.281668 +time/saving (s) 0.00343428 +time/training (s) 7.65035 +time/epoch (s) 10.598 +time/total (s) 774.356 +Epoch -929 +---------------------------------- --------------- +2022-05-10 13:23:39.036642 PDT | [1] Epoch -928 finished +---------------------------------- --------------- +epoch -928 +replay_buffer/size 999033 +trainer/num train calls 73000 +trainer/Policy Loss -19.4102 +trainer/Log Pis Mean 24.3444 +trainer/Log Pis Std 12.1775 +trainer/Log Pis Max 67.271 +trainer/Log Pis Min -12.0619 +trainer/policy/mean Mean -0.027648 +trainer/policy/mean Std 0.902485 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88683 +trainer/policy/normal/std Std 0.531108 +trainer/policy/normal/std Max 5.7827 +trainer/policy/normal/std Min 0.495198 +trainer/policy/normal/log_std Mean 1.0379 +trainer/policy/normal/log_std Std 0.230932 +trainer/policy/normal/log_std Max 1.75487 +trainer/policy/normal/log_std Min -0.702798 +eval/num steps total 73000 +eval/num paths total 73 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0648081 +eval/Actions Std 0.903671 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59405 +time/logging (s) 0.0038114 +time/sampling batch (s) 0.524876 +time/saving (s) 0.00338022 +time/training (s) 6.91038 +time/epoch (s) 10.0365 +time/total (s) 784.396 +Epoch -928 +---------------------------------- --------------- +2022-05-10 13:23:49.721265 PDT | [1] Epoch -927 finished +---------------------------------- --------------- +epoch -927 +replay_buffer/size 999033 +trainer/num train calls 74000 +trainer/Policy Loss -18.8846 +trainer/Log Pis Mean 23.5757 +trainer/Log Pis Std 13.5208 +trainer/Log Pis Max 75.2612 +trainer/Log Pis Min -6.54676 +trainer/policy/mean Mean -0.0186986 +trainer/policy/mean Std 0.902529 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.88671 +trainer/policy/normal/std Std 0.56743 +trainer/policy/normal/std Max 6.50516 +trainer/policy/normal/std Min 0.506946 +trainer/policy/normal/log_std Mean 1.03355 +trainer/policy/normal/log_std Std 0.254785 +trainer/policy/normal/log_std Max 1.8726 +trainer/policy/normal/log_std Min -0.67935 +eval/num steps total 74000 +eval/num paths total 74 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114595 +eval/Actions Std 0.909429 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.604 +time/logging (s) 0.00364363 +time/sampling batch (s) 0.273593 +time/saving (s) 0.00343003 +time/training (s) 7.78043 +time/epoch (s) 10.6651 +time/total (s) 795.064 +Epoch -927 +---------------------------------- --------------- +2022-05-10 13:23:59.871544 PDT | [1] Epoch -926 finished +---------------------------------- --------------- +epoch -926 +replay_buffer/size 999033 +trainer/num train calls 75000 +trainer/Policy Loss -18.4575 +trainer/Log Pis Mean 24.4825 +trainer/Log Pis Std 13.659 +trainer/Log Pis Max 70.4811 +trainer/Log Pis Min -13.1217 +trainer/policy/mean Mean -0.0440016 +trainer/policy/mean Std 0.899817 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91411 +trainer/policy/normal/std Std 0.545863 +trainer/policy/normal/std Max 5.7276 +trainer/policy/normal/std Min 0.53699 +trainer/policy/normal/log_std Mean 1.04667 +trainer/policy/normal/log_std Std 0.233575 +trainer/policy/normal/log_std Max 1.7453 +trainer/policy/normal/log_std Min -0.621776 +eval/num steps total 75000 +eval/num paths total 75 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0791082 +eval/Actions Std 0.909444 +eval/Actions Max 0.999987 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49172 +time/logging (s) 0.00397843 +time/sampling batch (s) 0.525038 +time/saving (s) 0.00418917 +time/training (s) 7.10615 +time/epoch (s) 10.1311 +time/total (s) 805.198 +Epoch -926 +---------------------------------- --------------- +2022-05-10 13:24:10.920975 PDT | [1] Epoch -925 finished +---------------------------------- --------------- +epoch -925 +replay_buffer/size 999033 +trainer/num train calls 76000 +trainer/Policy Loss -20.0519 +trainer/Log Pis Mean 24.1657 +trainer/Log Pis Std 12.7748 +trainer/Log Pis Max 59.6425 +trainer/Log Pis Min -15.7978 +trainer/policy/mean Mean -0.0274485 +trainer/policy/mean Std 0.90578 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91054 +trainer/policy/normal/std Std 0.532747 +trainer/policy/normal/std Max 4.56035 +trainer/policy/normal/std Min 0.480592 +trainer/policy/normal/log_std Mean 1.04563 +trainer/policy/normal/log_std Std 0.234926 +trainer/policy/normal/log_std Max 1.5174 +trainer/policy/normal/log_std Min -0.732737 +eval/num steps total 76000 +eval/num paths total 76 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0302013 +eval/Actions Std 0.937722 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51139 +time/logging (s) 0.00374659 +time/sampling batch (s) 1.02756 +time/saving (s) 0.00341521 +time/training (s) 7.48314 +time/epoch (s) 11.0292 +time/total (s) 816.23 +Epoch -925 +---------------------------------- --------------- +2022-05-10 13:24:22.282616 PDT | [1] Epoch -924 finished +---------------------------------- --------------- +epoch -924 +replay_buffer/size 999033 +trainer/num train calls 77000 +trainer/Policy Loss -19.0987 +trainer/Log Pis Mean 25.2863 +trainer/Log Pis Std 13.5305 +trainer/Log Pis Max 76.6656 +trainer/Log Pis Min -9.00157 +trainer/policy/mean Mean -0.0368023 +trainer/policy/mean Std 0.905269 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.90296 +trainer/policy/normal/std Std 0.544523 +trainer/policy/normal/std Max 5.23266 +trainer/policy/normal/std Min 0.462183 +trainer/policy/normal/log_std Mean 1.04204 +trainer/policy/normal/log_std Std 0.2401 +trainer/policy/normal/log_std Max 1.65492 +trainer/policy/normal/log_std Min -0.771795 +eval/num steps total 77000 +eval/num paths total 77 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0211704 +eval/Actions Std 0.924235 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63257 +time/logging (s) 0.00377371 +time/sampling batch (s) 0.276052 +time/saving (s) 0.00345411 +time/training (s) 8.42614 +time/epoch (s) 11.342 +time/total (s) 827.575 +Epoch -924 +---------------------------------- --------------- +2022-05-10 13:24:33.113810 PDT | [1] Epoch -923 finished +---------------------------------- --------------- +epoch -923 +replay_buffer/size 999033 +trainer/num train calls 78000 +trainer/Policy Loss -20.4264 +trainer/Log Pis Mean 23.5916 +trainer/Log Pis Std 12.8432 +trainer/Log Pis Max 80.1259 +trainer/Log Pis Min -6.41571 +trainer/policy/mean Mean -0.0367426 +trainer/policy/mean Std 0.902801 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87369 +trainer/policy/normal/std Std 0.546367 +trainer/policy/normal/std Max 5.4995 +trainer/policy/normal/std Min 0.45171 +trainer/policy/normal/log_std Mean 1.03154 +trainer/policy/normal/log_std Std 0.240815 +trainer/policy/normal/log_std Max 1.70466 +trainer/policy/normal/log_std Min -0.794714 +eval/num steps total 78000 +eval/num paths total 78 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.151836 +eval/Actions Std 0.878865 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47872 +time/logging (s) 0.00368921 +time/sampling batch (s) 0.777898 +time/saving (s) 0.00346768 +time/training (s) 7.54778 +time/epoch (s) 10.8116 +time/total (s) 838.389 +Epoch -923 +---------------------------------- --------------- +2022-05-10 13:24:43.802875 PDT | [1] Epoch -922 finished +---------------------------------- --------------- +epoch -922 +replay_buffer/size 999033 +trainer/num train calls 79000 +trainer/Policy Loss -17.9513 +trainer/Log Pis Mean 24.5814 +trainer/Log Pis Std 13.3495 +trainer/Log Pis Max 77.0647 +trainer/Log Pis Min -8.01849 +trainer/policy/mean Mean -0.0175258 +trainer/policy/mean Std 0.90679 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91043 +trainer/policy/normal/std Std 0.548896 +trainer/policy/normal/std Max 5.39629 +trainer/policy/normal/std Min 0.563127 +trainer/policy/normal/log_std Mean 1.04499 +trainer/policy/normal/log_std Std 0.235906 +trainer/policy/normal/log_std Max 1.68571 +trainer/policy/normal/log_std Min -0.574251 +eval/num steps total 79000 +eval/num paths total 79 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.217252 +eval/Actions Std 0.890824 +eval/Actions Max 0.999987 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57992 +time/logging (s) 0.00369906 +time/sampling batch (s) 0.775543 +time/saving (s) 0.00339073 +time/training (s) 7.30696 +time/epoch (s) 10.6695 +time/total (s) 849.062 +Epoch -922 +---------------------------------- --------------- +2022-05-10 13:24:55.498324 PDT | [1] Epoch -921 finished +---------------------------------- --------------- +epoch -921 +replay_buffer/size 999033 +trainer/num train calls 80000 +trainer/Policy Loss -18.6012 +trainer/Log Pis Mean 23.4776 +trainer/Log Pis Std 12.4413 +trainer/Log Pis Max 66.8901 +trainer/Log Pis Min -7.06153 +trainer/policy/mean Mean -0.0368574 +trainer/policy/mean Std 0.903211 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.83762 +trainer/policy/normal/std Std 0.538836 +trainer/policy/normal/std Max 4.89476 +trainer/policy/normal/std Min 0.412522 +trainer/policy/normal/log_std Mean 1.01884 +trainer/policy/normal/log_std Std 0.24125 +trainer/policy/normal/log_std Max 1.58817 +trainer/policy/normal/log_std Min -0.885466 +eval/num steps total 80000 +eval/num paths total 80 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.298909 +eval/Actions Std 0.89482 +eval/Actions Max 0.999969 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63393 +time/logging (s) 0.00408157 +time/sampling batch (s) 0.274404 +time/saving (s) 0.00392103 +time/training (s) 8.75994 +time/epoch (s) 11.6763 +time/total (s) 860.741 +Epoch -921 +---------------------------------- --------------- +2022-05-10 13:25:06.233477 PDT | [1] Epoch -920 finished +---------------------------------- --------------- +epoch -920 +replay_buffer/size 999033 +trainer/num train calls 81000 +trainer/Policy Loss -18.917 +trainer/Log Pis Mean 23.5551 +trainer/Log Pis Std 12.6067 +trainer/Log Pis Max 67.8086 +trainer/Log Pis Min -9.17781 +trainer/policy/mean Mean -0.0240391 +trainer/policy/mean Std 0.905151 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89701 +trainer/policy/normal/std Std 0.527829 +trainer/policy/normal/std Max 4.92945 +trainer/policy/normal/std Min 0.512254 +trainer/policy/normal/log_std Mean 1.04215 +trainer/policy/normal/log_std Std 0.225601 +trainer/policy/normal/log_std Max 1.59523 +trainer/policy/normal/log_std Min -0.668935 +eval/num steps total 81000 +eval/num paths total 81 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.379085 +eval/Actions Std 0.893015 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82077 +time/logging (s) 0.00374457 +time/sampling batch (s) 0.275074 +time/saving (s) 0.00352904 +time/training (s) 7.6118 +time/epoch (s) 10.7149 +time/total (s) 871.459 +Epoch -920 +---------------------------------- --------------- +2022-05-10 13:25:15.330200 PDT | [1] Epoch -919 finished +---------------------------------- --------------- +epoch -919 +replay_buffer/size 999033 +trainer/num train calls 82000 +trainer/Policy Loss -19.5249 +trainer/Log Pis Mean 24.558 +trainer/Log Pis Std 13.6981 +trainer/Log Pis Max 70.5967 +trainer/Log Pis Min -7.52537 +trainer/policy/mean Mean -0.0417663 +trainer/policy/mean Std 0.899924 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.9097 +trainer/policy/normal/std Std 0.533471 +trainer/policy/normal/std Max 5.6955 +trainer/policy/normal/std Min 0.65497 +trainer/policy/normal/log_std Mean 1.047 +trainer/policy/normal/log_std Std 0.221022 +trainer/policy/normal/log_std Max 1.73968 +trainer/policy/normal/log_std Min -0.423166 +eval/num steps total 82000 +eval/num paths total 82 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.163661 +eval/Actions Std 0.930836 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39822 +time/logging (s) 0.00433681 +time/sampling batch (s) 0.283462 +time/saving (s) 0.00425041 +time/training (s) 6.38683 +time/epoch (s) 9.0771 +time/total (s) 880.539 +Epoch -919 +---------------------------------- --------------- +2022-05-10 13:25:26.896675 PDT | [1] Epoch -918 finished +---------------------------------- --------------- +epoch -918 +replay_buffer/size 999033 +trainer/num train calls 83000 +trainer/Policy Loss -20.1278 +trainer/Log Pis Mean 24.5774 +trainer/Log Pis Std 12.8732 +trainer/Log Pis Max 74.1716 +trainer/Log Pis Min -6.23687 +trainer/policy/mean Mean -0.0143253 +trainer/policy/mean Std 0.899227 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86539 +trainer/policy/normal/std Std 0.550577 +trainer/policy/normal/std Max 5.1142 +trainer/policy/normal/std Min 0.492657 +trainer/policy/normal/log_std Mean 1.02719 +trainer/policy/normal/log_std Std 0.250663 +trainer/policy/normal/log_std Max 1.63202 +trainer/policy/normal/log_std Min -0.707941 +eval/num steps total 83000 +eval/num paths total 83 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0874854 +eval/Actions Std 0.901801 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66647 +time/logging (s) 0.00443827 +time/sampling batch (s) 0.287194 +time/saving (s) 0.00420183 +time/training (s) 8.58265 +time/epoch (s) 11.545 +time/total (s) 892.088 +Epoch -918 +---------------------------------- --------------- +2022-05-10 13:25:37.008564 PDT | [1] Epoch -917 finished +---------------------------------- --------------- +epoch -917 +replay_buffer/size 999033 +trainer/num train calls 84000 +trainer/Policy Loss -18.4294 +trainer/Log Pis Mean 23.7537 +trainer/Log Pis Std 12.9145 +trainer/Log Pis Max 72.1293 +trainer/Log Pis Min -5.60643 +trainer/policy/mean Mean -0.0486462 +trainer/policy/mean Std 0.897303 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.84459 +trainer/policy/normal/std Std 0.549609 +trainer/policy/normal/std Max 5.96958 +trainer/policy/normal/std Min 0.556695 +trainer/policy/normal/log_std Mean 1.0202 +trainer/policy/normal/log_std Std 0.247327 +trainer/policy/normal/log_std Max 1.78668 +trainer/policy/normal/log_std Min -0.585738 +eval/num steps total 84000 +eval/num paths total 84 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0698649 +eval/Actions Std 0.900243 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69103 +time/logging (s) 0.00403596 +time/sampling batch (s) 0.528821 +time/saving (s) 0.00397401 +time/training (s) 6.86323 +time/epoch (s) 10.0911 +time/total (s) 902.182 +Epoch -917 +---------------------------------- --------------- +2022-05-10 13:25:46.992694 PDT | [1] Epoch -916 finished +---------------------------------- --------------- +epoch -916 +replay_buffer/size 999033 +trainer/num train calls 85000 +trainer/Policy Loss -20.7612 +trainer/Log Pis Mean 23.2021 +trainer/Log Pis Std 12.7111 +trainer/Log Pis Max 70.7954 +trainer/Log Pis Min -6.24791 +trainer/policy/mean Mean -0.0272271 +trainer/policy/mean Std 0.902517 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87845 +trainer/policy/normal/std Std 0.536007 +trainer/policy/normal/std Max 4.66859 +trainer/policy/normal/std Min 0.507628 +trainer/policy/normal/log_std Mean 1.03421 +trainer/policy/normal/log_std Std 0.235361 +trainer/policy/normal/log_std Max 1.54086 +trainer/policy/normal/log_std Min -0.678007 +eval/num steps total 85000 +eval/num paths total 85 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00321875 +eval/Actions Std 0.900922 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63813 +time/logging (s) 0.00398566 +time/sampling batch (s) 0.281152 +time/saving (s) 0.00391046 +time/training (s) 7.03647 +time/epoch (s) 9.96364 +time/total (s) 912.149 +Epoch -916 +---------------------------------- --------------- +2022-05-10 13:25:58.035331 PDT | [1] Epoch -915 finished +---------------------------------- --------------- +epoch -915 +replay_buffer/size 999033 +trainer/num train calls 86000 +trainer/Policy Loss -19.8137 +trainer/Log Pis Mean 23.7698 +trainer/Log Pis Std 13.6502 +trainer/Log Pis Max 71.679 +trainer/Log Pis Min -7.89713 +trainer/policy/mean Mean -0.0414306 +trainer/policy/mean Std 0.90243 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83537 +trainer/policy/normal/std Std 0.532886 +trainer/policy/normal/std Max 5.10099 +trainer/policy/normal/std Min 0.517805 +trainer/policy/normal/log_std Mean 1.01913 +trainer/policy/normal/log_std Std 0.2342 +trainer/policy/normal/log_std Max 1.62944 +trainer/policy/normal/log_std Min -0.658156 +eval/num steps total 86000 +eval/num paths total 86 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00769845 +eval/Actions Std 0.964557 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56663 +time/logging (s) 0.00412785 +time/sampling batch (s) 0.528126 +time/saving (s) 0.00403517 +time/training (s) 7.91968 +time/epoch (s) 11.0226 +time/total (s) 923.175 +Epoch -915 +---------------------------------- --------------- +2022-05-10 13:26:08.500878 PDT | [1] Epoch -914 finished +---------------------------------- --------------- +epoch -914 +replay_buffer/size 999033 +trainer/num train calls 87000 +trainer/Policy Loss -19.3835 +trainer/Log Pis Mean 23.6691 +trainer/Log Pis Std 13.206 +trainer/Log Pis Max 69.3998 +trainer/Log Pis Min -7.52375 +trainer/policy/mean Mean -0.0227733 +trainer/policy/mean Std 0.903268 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.92288 +trainer/policy/normal/std Std 0.5488 +trainer/policy/normal/std Max 4.99687 +trainer/policy/normal/std Min 0.555767 +trainer/policy/normal/log_std Mean 1.0489 +trainer/policy/normal/log_std Std 0.23921 +trainer/policy/normal/log_std Max 1.60881 +trainer/policy/normal/log_std Min -0.587406 +eval/num steps total 87000 +eval/num paths total 87 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00521079 +eval/Actions Std 0.898212 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46156 +time/logging (s) 0.00376031 +time/sampling batch (s) 0.527831 +time/saving (s) 0.00354396 +time/training (s) 7.44815 +time/epoch (s) 10.4448 +time/total (s) 933.623 +Epoch -914 +---------------------------------- --------------- +2022-05-10 13:26:20.494918 PDT | [1] Epoch -913 finished +---------------------------------- --------------- +epoch -913 +replay_buffer/size 999033 +trainer/num train calls 88000 +trainer/Policy Loss -19.9171 +trainer/Log Pis Mean 24.6139 +trainer/Log Pis Std 12.7767 +trainer/Log Pis Max 75.2849 +trainer/Log Pis Min -3.56641 +trainer/policy/mean Mean -0.0396371 +trainer/policy/mean Std 0.901991 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.92305 +trainer/policy/normal/std Std 0.55487 +trainer/policy/normal/std Max 6.03334 +trainer/policy/normal/std Min 0.552836 +trainer/policy/normal/log_std Mean 1.04936 +trainer/policy/normal/log_std Std 0.23478 +trainer/policy/normal/log_std Max 1.7973 +trainer/policy/normal/log_std Min -0.592693 +eval/num steps total 88000 +eval/num paths total 88 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.331145 +eval/Actions Std 0.768809 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50018 +time/logging (s) 0.00377838 +time/sampling batch (s) 0.534419 +time/saving (s) 0.00342753 +time/training (s) 8.93191 +time/epoch (s) 11.9737 +time/total (s) 945.6 +Epoch -913 +---------------------------------- --------------- +2022-05-10 13:26:31.810692 PDT | [1] Epoch -912 finished +---------------------------------- --------------- +epoch -912 +replay_buffer/size 999033 +trainer/num train calls 89000 +trainer/Policy Loss -19.6933 +trainer/Log Pis Mean 25.0258 +trainer/Log Pis Std 13.6967 +trainer/Log Pis Max 64.7876 +trainer/Log Pis Min -12.6217 +trainer/policy/mean Mean -0.0462923 +trainer/policy/mean Std 0.904297 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89401 +trainer/policy/normal/std Std 0.534775 +trainer/policy/normal/std Max 4.86075 +trainer/policy/normal/std Min 0.522713 +trainer/policy/normal/log_std Mean 1.04016 +trainer/policy/normal/log_std Std 0.231992 +trainer/policy/normal/log_std Max 1.58119 +trainer/policy/normal/log_std Min -0.648723 +eval/num steps total 89000 +eval/num paths total 89 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0930524 +eval/Actions Std 0.906169 +eval/Actions Max 0.999983 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59511 +time/logging (s) 0.00506649 +time/sampling batch (s) 0.783015 +time/saving (s) 0.004116 +time/training (s) 7.9096 +time/epoch (s) 11.2969 +time/total (s) 956.9 +Epoch -912 +---------------------------------- --------------- +2022-05-10 13:26:41.822387 PDT | [1] Epoch -911 finished +---------------------------------- --------------- +epoch -911 +replay_buffer/size 999033 +trainer/num train calls 90000 +trainer/Policy Loss -19.1787 +trainer/Log Pis Mean 23.97 +trainer/Log Pis Std 12.6532 +trainer/Log Pis Max 68.6285 +trainer/Log Pis Min -4.9662 +trainer/policy/mean Mean -0.0288085 +trainer/policy/mean Std 0.902609 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87805 +trainer/policy/normal/std Std 0.569221 +trainer/policy/normal/std Max 4.93396 +trainer/policy/normal/std Min 0.492125 +trainer/policy/normal/log_std Mean 1.03059 +trainer/policy/normal/log_std Std 0.253669 +trainer/policy/normal/log_std Max 1.59614 +trainer/policy/normal/log_std Min -0.709022 +eval/num steps total 90000 +eval/num paths total 90 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.191855 +eval/Actions Std 0.872295 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64826 +time/logging (s) 0.00372609 +time/sampling batch (s) 0.275806 +time/saving (s) 0.00345953 +time/training (s) 7.05816 +time/epoch (s) 9.98941 +time/total (s) 966.893 +Epoch -911 +---------------------------------- --------------- +2022-05-10 13:26:51.544489 PDT | [1] Epoch -910 finished +---------------------------------- --------------- +epoch -910 +replay_buffer/size 999033 +trainer/num train calls 91000 +trainer/Policy Loss -18.3324 +trainer/Log Pis Mean 23.4025 +trainer/Log Pis Std 13.2774 +trainer/Log Pis Max 65.4975 +trainer/Log Pis Min -7.97379 +trainer/policy/mean Mean -0.0539649 +trainer/policy/mean Std 0.9013 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.91135 +trainer/policy/normal/std Std 0.571202 +trainer/policy/normal/std Max 4.82109 +trainer/policy/normal/std Min 0.502406 +trainer/policy/normal/log_std Mean 1.04219 +trainer/policy/normal/log_std Std 0.254736 +trainer/policy/normal/log_std Max 1.573 +trainer/policy/normal/log_std Min -0.688347 +eval/num steps total 91000 +eval/num paths total 91 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.1524 +eval/Actions Std 0.907885 +eval/Actions Max 0.999982 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.736 +time/logging (s) 0.0037916 +time/sampling batch (s) 0.274685 +time/saving (s) 0.00346147 +time/training (s) 6.68477 +time/epoch (s) 9.70271 +time/total (s) 976.599 +Epoch -910 +---------------------------------- --------------- +2022-05-10 13:27:02.119380 PDT | [1] Epoch -909 finished +---------------------------------- --------------- +epoch -909 +replay_buffer/size 999033 +trainer/num train calls 92000 +trainer/Policy Loss -19.0898 +trainer/Log Pis Mean 23.6118 +trainer/Log Pis Std 12.6851 +trainer/Log Pis Max 74.3999 +trainer/Log Pis Min -7.73672 +trainer/policy/mean Mean -0.0198566 +trainer/policy/mean Std 0.90472 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91758 +trainer/policy/normal/std Std 0.567392 +trainer/policy/normal/std Max 6.08238 +trainer/policy/normal/std Min 0.542666 +trainer/policy/normal/log_std Mean 1.04583 +trainer/policy/normal/log_std Std 0.244552 +trainer/policy/normal/log_std Max 1.8054 +trainer/policy/normal/log_std Min -0.611261 +eval/num steps total 92000 +eval/num paths total 92 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0441608 +eval/Actions Std 0.908593 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54078 +time/logging (s) 0.00369271 +time/sampling batch (s) 0.276651 +time/saving (s) 0.00343409 +time/training (s) 7.73056 +time/epoch (s) 10.5551 +time/total (s) 987.157 +Epoch -909 +---------------------------------- --------------- +2022-05-10 13:27:12.777790 PDT | [1] Epoch -908 finished +---------------------------------- --------------- +epoch -908 +replay_buffer/size 999033 +trainer/num train calls 93000 +trainer/Policy Loss -18.109 +trainer/Log Pis Mean 24.6324 +trainer/Log Pis Std 13.3459 +trainer/Log Pis Max 70.3936 +trainer/Log Pis Min -15.0986 +trainer/policy/mean Mean -0.0292208 +trainer/policy/mean Std 0.900723 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.952 +trainer/policy/normal/std Std 0.584583 +trainer/policy/normal/std Max 5.79807 +trainer/policy/normal/std Min 0.534452 +trainer/policy/normal/log_std Mean 1.05674 +trainer/policy/normal/log_std Std 0.248016 +trainer/policy/normal/log_std Max 1.75753 +trainer/policy/normal/log_std Min -0.626514 +eval/num steps total 93000 +eval/num paths total 93 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.182387 +eval/Actions Std 0.852595 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53789 +time/logging (s) 0.00379168 +time/sampling batch (s) 0.282073 +time/saving (s) 0.00344164 +time/training (s) 7.81138 +time/epoch (s) 10.6386 +time/total (s) 997.798 +Epoch -908 +---------------------------------- --------------- +2022-05-10 13:27:23.801519 PDT | [1] Epoch -907 finished +---------------------------------- --------------- +epoch -907 +replay_buffer/size 999033 +trainer/num train calls 94000 +trainer/Policy Loss -19.1144 +trainer/Log Pis Mean 24.1883 +trainer/Log Pis Std 12.7828 +trainer/Log Pis Max 70.7385 +trainer/Log Pis Min -9.4171 +trainer/policy/mean Mean -0.0288167 +trainer/policy/mean Std 0.900206 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84406 +trainer/policy/normal/std Std 0.534354 +trainer/policy/normal/std Max 5.29276 +trainer/policy/normal/std Min 0.462974 +trainer/policy/normal/log_std Mean 1.02154 +trainer/policy/normal/log_std Std 0.239652 +trainer/policy/normal/log_std Max 1.66634 +trainer/policy/normal/log_std Min -0.770084 +eval/num steps total 94000 +eval/num paths total 94 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.363952 +eval/Actions Std 0.791999 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45236 +time/logging (s) 0.00423607 +time/sampling batch (s) 1.02716 +time/saving (s) 0.00385757 +time/training (s) 7.51686 +time/epoch (s) 11.0045 +time/total (s) 1008.81 +Epoch -907 +---------------------------------- --------------- +2022-05-10 13:27:34.185211 PDT | [1] Epoch -906 finished +---------------------------------- --------------- +epoch -906 +replay_buffer/size 999033 +trainer/num train calls 95000 +trainer/Policy Loss -18.9796 +trainer/Log Pis Mean 24.8267 +trainer/Log Pis Std 13.5564 +trainer/Log Pis Max 74.1016 +trainer/Log Pis Min -6.46149 +trainer/policy/mean Mean -0.0426071 +trainer/policy/mean Std 0.90192 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90362 +trainer/policy/normal/std Std 0.558089 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.530029 +trainer/policy/normal/log_std Mean 1.04173 +trainer/policy/normal/log_std Std 0.240817 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.634824 +eval/num steps total 95000 +eval/num paths total 95 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109462 +eval/Actions Std 0.845872 +eval/Actions Max 0.999997 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58864 +time/logging (s) 0.00368702 +time/sampling batch (s) 0.527006 +time/saving (s) 0.00355686 +time/training (s) 7.23998 +time/epoch (s) 10.3629 +time/total (s) 1019.17 +Epoch -906 +---------------------------------- --------------- +2022-05-10 13:27:44.735688 PDT | [1] Epoch -905 finished +---------------------------------- --------------- +epoch -905 +replay_buffer/size 999033 +trainer/num train calls 96000 +trainer/Policy Loss -18.8703 +trainer/Log Pis Mean 24.9763 +trainer/Log Pis Std 13.8162 +trainer/Log Pis Max 73.1172 +trainer/Log Pis Min -7.02918 +trainer/policy/mean Mean -0.0321525 +trainer/policy/mean Std 0.904703 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.91746 +trainer/policy/normal/std Std 0.553551 +trainer/policy/normal/std Max 5.35881 +trainer/policy/normal/std Min 0.510558 +trainer/policy/normal/log_std Mean 1.04694 +trainer/policy/normal/log_std Std 0.238813 +trainer/policy/normal/log_std Max 1.67874 +trainer/policy/normal/log_std Min -0.672251 +eval/num steps total 96000 +eval/num paths total 96 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0885561 +eval/Actions Std 0.911342 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53363 +time/logging (s) 0.00381059 +time/sampling batch (s) 0.278758 +time/saving (s) 0.00354784 +time/training (s) 7.71076 +time/epoch (s) 10.5305 +time/total (s) 1029.7 +Epoch -905 +---------------------------------- --------------- +2022-05-10 13:27:56.034654 PDT | [1] Epoch -904 finished +---------------------------------- --------------- +epoch -904 +replay_buffer/size 999033 +trainer/num train calls 97000 +trainer/Policy Loss -20.2989 +trainer/Log Pis Mean 24.0229 +trainer/Log Pis Std 13.3959 +trainer/Log Pis Max 62.6486 +trainer/Log Pis Min -5.62282 +trainer/policy/mean Mean -0.0465686 +trainer/policy/mean Std 0.899905 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87401 +trainer/policy/normal/std Std 0.545881 +trainer/policy/normal/std Max 7.06774 +trainer/policy/normal/std Min 0.48829 +trainer/policy/normal/log_std Mean 1.03188 +trainer/policy/normal/log_std Std 0.238629 +trainer/policy/normal/log_std Max 1.95554 +trainer/policy/normal/log_std Min -0.716847 +eval/num steps total 97000 +eval/num paths total 97 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.161827 +eval/Actions Std 0.925595 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53588 +time/logging (s) 0.0037217 +time/sampling batch (s) 1.0275 +time/saving (s) 0.00344485 +time/training (s) 7.7084 +time/epoch (s) 11.279 +time/total (s) 1040.99 +Epoch -904 +---------------------------------- --------------- +2022-05-10 13:28:06.984942 PDT | [1] Epoch -903 finished +---------------------------------- --------------- +epoch -903 +replay_buffer/size 999033 +trainer/num train calls 98000 +trainer/Policy Loss -19.65 +trainer/Log Pis Mean 24.2327 +trainer/Log Pis Std 12.8725 +trainer/Log Pis Max 63.6559 +trainer/Log Pis Min -7.10219 +trainer/policy/mean Mean -0.0430721 +trainer/policy/mean Std 0.905559 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82727 +trainer/policy/normal/std Std 0.558066 +trainer/policy/normal/std Max 5.596 +trainer/policy/normal/std Min 0.521266 +trainer/policy/normal/log_std Mean 1.01273 +trainer/policy/normal/log_std Std 0.254752 +trainer/policy/normal/log_std Max 1.72205 +trainer/policy/normal/log_std Min -0.651495 +eval/num steps total 98000 +eval/num paths total 98 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.224491 +eval/Actions Std 0.945736 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58365 +time/logging (s) 0.00376334 +time/sampling batch (s) 0.531481 +time/saving (s) 0.00355635 +time/training (s) 7.8076 +time/epoch (s) 10.93 +time/total (s) 1051.92 +Epoch -903 +---------------------------------- --------------- +2022-05-10 13:28:17.240734 PDT | [1] Epoch -902 finished +---------------------------------- --------------- +epoch -902 +replay_buffer/size 999033 +trainer/num train calls 99000 +trainer/Policy Loss -19.6869 +trainer/Log Pis Mean 24.6015 +trainer/Log Pis Std 13.3418 +trainer/Log Pis Max 64.5403 +trainer/Log Pis Min -8.74138 +trainer/policy/mean Mean -0.0531292 +trainer/policy/mean Std 0.9025 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.93056 +trainer/policy/normal/std Std 0.566575 +trainer/policy/normal/std Max 5.50949 +trainer/policy/normal/std Min 0.490577 +trainer/policy/normal/log_std Mean 1.05045 +trainer/policy/normal/log_std Std 0.244275 +trainer/policy/normal/log_std Max 1.70647 +trainer/policy/normal/log_std Min -0.712172 +eval/num steps total 99000 +eval/num paths total 99 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.282081 +eval/Actions Std 0.876235 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63006 +time/logging (s) 0.00414268 +time/sampling batch (s) 0.282135 +time/saving (s) 0.0040017 +time/training (s) 7.31562 +time/epoch (s) 10.236 +time/total (s) 1062.16 +Epoch -902 +---------------------------------- --------------- +2022-05-10 13:28:27.537898 PDT | [1] Epoch -901 finished +---------------------------------- --------------- +epoch -901 +replay_buffer/size 999033 +trainer/num train calls 100000 +trainer/Policy Loss -19.1337 +trainer/Log Pis Mean 23.9286 +trainer/Log Pis Std 12.8362 +trainer/Log Pis Max 63.2928 +trainer/Log Pis Min -5.97447 +trainer/policy/mean Mean -0.0338864 +trainer/policy/mean Std 0.907668 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84453 +trainer/policy/normal/std Std 0.56175 +trainer/policy/normal/std Max 5.4587 +trainer/policy/normal/std Min 0.454185 +trainer/policy/normal/log_std Mean 1.01943 +trainer/policy/normal/log_std Std 0.250598 +trainer/policy/normal/log_std Max 1.69721 +trainer/policy/normal/log_std Min -0.78925 +eval/num steps total 100000 +eval/num paths total 100 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138633 +eval/Actions Std 0.901089 +eval/Actions Max 0.999985 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70632 +time/logging (s) 0.00372386 +time/sampling batch (s) 0.525366 +time/saving (s) 0.0064359 +time/training (s) 7.03464 +time/epoch (s) 10.2765 +time/total (s) 1072.44 +Epoch -901 +---------------------------------- --------------- +2022-05-10 13:28:38.882147 PDT | [1] Epoch -900 finished +---------------------------------- --------------- +epoch -900 +replay_buffer/size 999033 +trainer/num train calls 101000 +trainer/Policy Loss -19.8885 +trainer/Log Pis Mean 25.0529 +trainer/Log Pis Std 13.4398 +trainer/Log Pis Max 76.7734 +trainer/Log Pis Min -9.25946 +trainer/policy/mean Mean -0.0388794 +trainer/policy/mean Std 0.903049 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.90476 +trainer/policy/normal/std Std 0.555889 +trainer/policy/normal/std Max 5.02017 +trainer/policy/normal/std Min 0.487151 +trainer/policy/normal/log_std Mean 1.04203 +trainer/policy/normal/log_std Std 0.241885 +trainer/policy/normal/log_std Max 1.61346 +trainer/policy/normal/log_std Min -0.719181 +eval/num steps total 101000 +eval/num paths total 101 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.205451 +eval/Actions Std 0.942544 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68413 +time/logging (s) 0.00365806 +time/sampling batch (s) 1.02552 +time/saving (s) 0.00340814 +time/training (s) 7.60776 +time/epoch (s) 11.3245 +time/total (s) 1083.77 +Epoch -900 +---------------------------------- --------------- +2022-05-10 13:28:48.844464 PDT | [1] Epoch -899 finished +---------------------------------- --------------- +epoch -899 +replay_buffer/size 999033 +trainer/num train calls 102000 +trainer/Policy Loss -19.498 +trainer/Log Pis Mean 24.279 +trainer/Log Pis Std 12.8366 +trainer/Log Pis Max 72.087 +trainer/Log Pis Min -5.77307 +trainer/policy/mean Mean -0.0384023 +trainer/policy/mean Std 0.903363 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84274 +trainer/policy/normal/std Std 0.565524 +trainer/policy/normal/std Max 5.80227 +trainer/policy/normal/std Min 0.500429 +trainer/policy/normal/log_std Mean 1.01721 +trainer/policy/normal/log_std Std 0.26112 +trainer/policy/normal/log_std Max 1.75825 +trainer/policy/normal/log_std Min -0.692289 +eval/num steps total 102000 +eval/num paths total 102 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0516691 +eval/Actions Std 0.904381 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39685 +time/logging (s) 0.00371847 +time/sampling batch (s) 0.277144 +time/saving (s) 0.00345066 +time/training (s) 7.26134 +time/epoch (s) 9.9425 +time/total (s) 1093.71 +Epoch -899 +---------------------------------- --------------- +2022-05-10 13:29:00.146477 PDT | [1] Epoch -898 finished +---------------------------------- --------------- +epoch -898 +replay_buffer/size 999033 +trainer/num train calls 103000 +trainer/Policy Loss -19.0721 +trainer/Log Pis Mean 24.5748 +trainer/Log Pis Std 12.732 +trainer/Log Pis Max 70.3686 +trainer/Log Pis Min -11.6039 +trainer/policy/mean Mean -0.0609719 +trainer/policy/mean Std 0.900044 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82802 +trainer/policy/normal/std Std 0.555186 +trainer/policy/normal/std Max 5.54334 +trainer/policy/normal/std Min 0.472656 +trainer/policy/normal/log_std Mean 1.01345 +trainer/policy/normal/log_std Std 0.252073 +trainer/policy/normal/log_std Max 1.7126 +trainer/policy/normal/log_std Min -0.749387 +eval/num steps total 103000 +eval/num paths total 103 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114857 +eval/Actions Std 0.885704 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60444 +time/logging (s) 0.00366509 +time/sampling batch (s) 0.5252 +time/saving (s) 0.00341634 +time/training (s) 8.14572 +time/epoch (s) 11.2824 +time/total (s) 1105 +Epoch -898 +---------------------------------- --------------- +2022-05-10 13:29:10.837817 PDT | [1] Epoch -897 finished +---------------------------------- --------------- +epoch -897 +replay_buffer/size 999033 +trainer/num train calls 104000 +trainer/Policy Loss -18.4038 +trainer/Log Pis Mean 24.411 +trainer/Log Pis Std 13.181 +trainer/Log Pis Max 73.6333 +trainer/Log Pis Min -9.80753 +trainer/policy/mean Mean -0.0402258 +trainer/policy/mean Std 0.902175 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.89403 +trainer/policy/normal/std Std 0.580668 +trainer/policy/normal/std Max 5.04002 +trainer/policy/normal/std Min 0.468067 +trainer/policy/normal/log_std Mean 1.03483 +trainer/policy/normal/log_std Std 0.26165 +trainer/policy/normal/log_std Max 1.61741 +trainer/policy/normal/log_std Min -0.759144 +eval/num steps total 104000 +eval/num paths total 104 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0805935 +eval/Actions Std 0.923524 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56072 +time/logging (s) 0.0040377 +time/sampling batch (s) 1.02521 +time/saving (s) 0.00380949 +time/training (s) 7.07806 +time/epoch (s) 10.6718 +time/total (s) 1115.67 +Epoch -897 +---------------------------------- --------------- +2022-05-10 13:29:22.077660 PDT | [1] Epoch -896 finished +---------------------------------- --------------- +epoch -896 +replay_buffer/size 999033 +trainer/num train calls 105000 +trainer/Policy Loss -19.5579 +trainer/Log Pis Mean 24.436 +trainer/Log Pis Std 13.7714 +trainer/Log Pis Max 74.5565 +trainer/Log Pis Min -11.8407 +trainer/policy/mean Mean -0.0325851 +trainer/policy/mean Std 0.89993 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83909 +trainer/policy/normal/std Std 0.560432 +trainer/policy/normal/std Max 4.49489 +trainer/policy/normal/std Min 0.453984 +trainer/policy/normal/log_std Mean 1.0165 +trainer/policy/normal/log_std Std 0.25729 +trainer/policy/normal/log_std Max 1.50294 +trainer/policy/normal/log_std Min -0.789694 +eval/num steps total 105000 +eval/num paths total 105 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0110752 +eval/Actions Std 0.937915 +eval/Actions Max 0.999996 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68689 +time/logging (s) 0.00381906 +time/sampling batch (s) 0.776014 +time/saving (s) 0.00371323 +time/training (s) 7.74907 +time/epoch (s) 11.2195 +time/total (s) 1126.89 +Epoch -896 +---------------------------------- --------------- +2022-05-10 13:29:32.503690 PDT | [1] Epoch -895 finished +---------------------------------- --------------- +epoch -895 +replay_buffer/size 999033 +trainer/num train calls 106000 +trainer/Policy Loss -19.2984 +trainer/Log Pis Mean 24.7881 +trainer/Log Pis Std 12.9385 +trainer/Log Pis Max 71.368 +trainer/Log Pis Min -9.39265 +trainer/policy/mean Mean -0.0365569 +trainer/policy/mean Std 0.907975 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85683 +trainer/policy/normal/std Std 0.531757 +trainer/policy/normal/std Max 4.79818 +trainer/policy/normal/std Min 0.486006 +trainer/policy/normal/log_std Mean 1.02636 +trainer/policy/normal/log_std Std 0.237466 +trainer/policy/normal/log_std Max 1.56824 +trainer/policy/normal/log_std Min -0.721535 +eval/num steps total 106000 +eval/num paths total 106 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.562659 +eval/Actions Std 0.772852 +eval/Actions Max 0.999992 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63752 +time/logging (s) 0.00393931 +time/sampling batch (s) 0.775679 +time/saving (s) 0.00357417 +time/training (s) 6.98552 +time/epoch (s) 10.4062 +time/total (s) 1137.3 +Epoch -895 +---------------------------------- --------------- +2022-05-10 13:29:42.667496 PDT | [1] Epoch -894 finished +---------------------------------- --------------- +epoch -894 +replay_buffer/size 999033 +trainer/num train calls 107000 +trainer/Policy Loss -20.4605 +trainer/Log Pis Mean 24.6358 +trainer/Log Pis Std 12.8817 +trainer/Log Pis Max 62.0343 +trainer/Log Pis Min -6.75577 +trainer/policy/mean Mean -0.0331977 +trainer/policy/mean Std 0.902059 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91532 +trainer/policy/normal/std Std 0.556867 +trainer/policy/normal/std Max 5.71517 +trainer/policy/normal/std Min 0.524891 +trainer/policy/normal/log_std Mean 1.04599 +trainer/policy/normal/log_std Std 0.239948 +trainer/policy/normal/log_std Max 1.74312 +trainer/policy/normal/log_std Min -0.644566 +eval/num steps total 106778 +eval/num paths total 107 +eval/path length Mean 778 +eval/path length Std 0 +eval/path length Max 778 +eval/path length Min 778 +eval/Rewards Mean 0.00128535 +eval/Rewards Std 0.0358287 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0103622 +eval/Actions Std 0.893317 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.57667 +time/logging (s) 0.00319908 +time/sampling batch (s) 0.526283 +time/saving (s) 0.00342489 +time/training (s) 7.03362 +time/epoch (s) 10.1432 +time/total (s) 1147.45 +Epoch -894 +---------------------------------- --------------- +2022-05-10 13:29:52.079282 PDT | [1] Epoch -893 finished +---------------------------------- --------------- +epoch -893 +replay_buffer/size 999033 +trainer/num train calls 108000 +trainer/Policy Loss -19.653 +trainer/Log Pis Mean 24.7269 +trainer/Log Pis Std 12.5237 +trainer/Log Pis Max 62.0503 +trainer/Log Pis Min -7.94041 +trainer/policy/mean Mean -0.0544617 +trainer/policy/mean Std 0.901576 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85672 +trainer/policy/normal/std Std 0.574891 +trainer/policy/normal/std Max 5.07819 +trainer/policy/normal/std Min 0.451478 +trainer/policy/normal/log_std Mean 1.02196 +trainer/policy/normal/log_std Std 0.260356 +trainer/policy/normal/log_std Max 1.62495 +trainer/policy/normal/log_std Min -0.795229 +eval/num steps total 107778 +eval/num paths total 108 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118336 +eval/Actions Std 0.913221 +eval/Actions Max 0.999985 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51077 +time/logging (s) 0.00371564 +time/sampling batch (s) 0.282544 +time/saving (s) 0.00342666 +time/training (s) 6.59187 +time/epoch (s) 9.39232 +time/total (s) 1156.84 +Epoch -893 +---------------------------------- --------------- +2022-05-10 13:30:02.051861 PDT | [1] Epoch -892 finished +---------------------------------- --------------- +epoch -892 +replay_buffer/size 999033 +trainer/num train calls 109000 +trainer/Policy Loss -20.3777 +trainer/Log Pis Mean 23.7986 +trainer/Log Pis Std 12.5242 +trainer/Log Pis Max 69.492 +trainer/Log Pis Min -3.95143 +trainer/policy/mean Mean -0.0616779 +trainer/policy/mean Std 0.902216 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83005 +trainer/policy/normal/std Std 0.565722 +trainer/policy/normal/std Max 5.70093 +trainer/policy/normal/std Min 0.492686 +trainer/policy/normal/log_std Mean 1.01264 +trainer/policy/normal/log_std Std 0.261065 +trainer/policy/normal/log_std Max 1.74063 +trainer/policy/normal/log_std Min -0.707883 +eval/num steps total 108778 +eval/num paths total 109 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0109127 +eval/Actions Std 0.894357 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34897 +time/logging (s) 0.00375315 +time/sampling batch (s) 0.525106 +time/saving (s) 0.0034406 +time/training (s) 7.07173 +time/epoch (s) 9.95299 +time/total (s) 1166.8 +Epoch -892 +---------------------------------- --------------- +2022-05-10 13:30:12.320612 PDT | [1] Epoch -891 finished +---------------------------------- --------------- +epoch -891 +replay_buffer/size 999033 +trainer/num train calls 110000 +trainer/Policy Loss -18.36 +trainer/Log Pis Mean 23.8509 +trainer/Log Pis Std 13.4817 +trainer/Log Pis Max 61.7891 +trainer/Log Pis Min -8.33278 +trainer/policy/mean Mean -0.0292884 +trainer/policy/mean Std 0.899695 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81351 +trainer/policy/normal/std Std 0.584029 +trainer/policy/normal/std Max 4.98672 +trainer/policy/normal/std Min 0.435488 +trainer/policy/normal/log_std Mean 1.00459 +trainer/policy/normal/log_std Std 0.27097 +trainer/policy/normal/log_std Max 1.60678 +trainer/policy/normal/log_std Min -0.831289 +eval/num steps total 109778 +eval/num paths total 110 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.21282 +eval/Actions Std 0.873309 +eval/Actions Max 0.999979 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74815 +time/logging (s) 0.00399518 +time/sampling batch (s) 0.275166 +time/saving (s) 0.00383616 +time/training (s) 7.21809 +time/epoch (s) 10.2492 +time/total (s) 1177.05 +Epoch -891 +---------------------------------- --------------- +2022-05-10 13:30:22.661303 PDT | [1] Epoch -890 finished +---------------------------------- --------------- +epoch -890 +replay_buffer/size 999033 +trainer/num train calls 111000 +trainer/Policy Loss -19.0792 +trainer/Log Pis Mean 24.4358 +trainer/Log Pis Std 14.1504 +trainer/Log Pis Max 80.2646 +trainer/Log Pis Min -12.6014 +trainer/policy/mean Mean -0.0102769 +trainer/policy/mean Std 0.901744 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.9338 +trainer/policy/normal/std Std 0.586675 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.459187 +trainer/policy/normal/log_std Mean 1.04937 +trainer/policy/normal/log_std Std 0.255895 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.778298 +eval/num steps total 110778 +eval/num paths total 111 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.373185 +eval/Actions Std 0.837313 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53929 +time/logging (s) 0.00394846 +time/sampling batch (s) 0.526993 +time/saving (s) 0.00363932 +time/training (s) 7.24669 +time/epoch (s) 10.3206 +time/total (s) 1187.38 +Epoch -890 +---------------------------------- --------------- +2022-05-10 13:30:33.174071 PDT | [1] Epoch -889 finished +---------------------------------- --------------- +epoch -889 +replay_buffer/size 999033 +trainer/num train calls 112000 +trainer/Policy Loss -20.6561 +trainer/Log Pis Mean 24.8984 +trainer/Log Pis Std 12.8294 +trainer/Log Pis Max 68.026 +trainer/Log Pis Min -4.21309 +trainer/policy/mean Mean -0.0315735 +trainer/policy/mean Std 0.905549 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.88214 +trainer/policy/normal/std Std 0.590171 +trainer/policy/normal/std Max 4.97505 +trainer/policy/normal/std Min 0.43101 +trainer/policy/normal/log_std Mean 1.02901 +trainer/policy/normal/log_std Std 0.27093 +trainer/policy/normal/log_std Max 1.60444 +trainer/policy/normal/log_std Min -0.841625 +eval/num steps total 111778 +eval/num paths total 112 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.169645 +eval/Actions Std 0.884391 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54526 +time/logging (s) 0.00377644 +time/sampling batch (s) 0.524933 +time/saving (s) 0.00342798 +time/training (s) 7.4153 +time/epoch (s) 10.4927 +time/total (s) 1197.87 +Epoch -889 +---------------------------------- --------------- +2022-05-10 13:30:43.215487 PDT | [1] Epoch -888 finished +---------------------------------- --------------- +epoch -888 +replay_buffer/size 999033 +trainer/num train calls 113000 +trainer/Policy Loss -19.8786 +trainer/Log Pis Mean 24.5547 +trainer/Log Pis Std 13.1778 +trainer/Log Pis Max 92.3485 +trainer/Log Pis Min -2.22261 +trainer/policy/mean Mean -0.0267771 +trainer/policy/mean Std 0.903935 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.90525 +trainer/policy/normal/std Std 0.56179 +trainer/policy/normal/std Max 5.31579 +trainer/policy/normal/std Min 0.420857 +trainer/policy/normal/log_std Mean 1.04102 +trainer/policy/normal/log_std Std 0.250216 +trainer/policy/normal/log_std Max 1.67068 +trainer/policy/normal/log_std Min -0.865463 +eval/num steps total 112778 +eval/num paths total 113 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.186024 +eval/Actions Std 0.90908 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42302 +time/logging (s) 0.00375593 +time/sampling batch (s) 0.526974 +time/saving (s) 0.003583 +time/training (s) 7.06396 +time/epoch (s) 10.0213 +time/total (s) 1207.9 +Epoch -888 +---------------------------------- --------------- +2022-05-10 13:30:53.373260 PDT | [1] Epoch -887 finished +---------------------------------- --------------- +epoch -887 +replay_buffer/size 999033 +trainer/num train calls 114000 +trainer/Policy Loss -19.841 +trainer/Log Pis Mean 24.7607 +trainer/Log Pis Std 13.8322 +trainer/Log Pis Max 67.7018 +trainer/Log Pis Min -7.67286 +trainer/policy/mean Mean -0.0387637 +trainer/policy/mean Std 0.904719 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.86152 +trainer/policy/normal/std Std 0.570405 +trainer/policy/normal/std Max 5.60343 +trainer/policy/normal/std Min 0.504733 +trainer/policy/normal/log_std Mean 1.02354 +trainer/policy/normal/log_std Std 0.262775 +trainer/policy/normal/log_std Max 1.72338 +trainer/policy/normal/log_std Min -0.683726 +eval/num steps total 113778 +eval/num paths total 114 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108551 +eval/Actions Std 0.905217 +eval/Actions Max 0.999995 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52408 +time/logging (s) 0.00421539 +time/sampling batch (s) 0.289049 +time/saving (s) 0.00431202 +time/training (s) 7.31598 +time/epoch (s) 10.1376 +time/total (s) 1218.04 +Epoch -887 +---------------------------------- --------------- +2022-05-10 13:31:04.171787 PDT | [1] Epoch -886 finished +---------------------------------- --------------- +epoch -886 +replay_buffer/size 999033 +trainer/num train calls 115000 +trainer/Policy Loss -19.2297 +trainer/Log Pis Mean 24.2681 +trainer/Log Pis Std 13.5417 +trainer/Log Pis Max 69.1744 +trainer/Log Pis Min -7.78841 +trainer/policy/mean Mean -0.0353989 +trainer/policy/mean Std 0.899904 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81067 +trainer/policy/normal/std Std 0.550395 +trainer/policy/normal/std Max 4.6605 +trainer/policy/normal/std Min 0.443852 +trainer/policy/normal/log_std Mean 1.0066 +trainer/policy/normal/log_std Std 0.257577 +trainer/policy/normal/log_std Max 1.53912 +trainer/policy/normal/log_std Min -0.812265 +eval/num steps total 114778 +eval/num paths total 115 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.322743 +eval/Actions Std 0.853608 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62164 +time/logging (s) 0.00419156 +time/sampling batch (s) 0.283692 +time/saving (s) 0.00413558 +time/training (s) 7.86359 +time/epoch (s) 10.7772 +time/total (s) 1228.82 +Epoch -886 +---------------------------------- --------------- +2022-05-10 13:31:15.064601 PDT | [1] Epoch -885 finished +---------------------------------- --------------- +epoch -885 +replay_buffer/size 999033 +trainer/num train calls 116000 +trainer/Policy Loss -19.9436 +trainer/Log Pis Mean 24.0111 +trainer/Log Pis Std 13.0952 +trainer/Log Pis Max 63.3961 +trainer/Log Pis Min -8.11843 +trainer/policy/mean Mean -0.045602 +trainer/policy/mean Std 0.908162 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.84763 +trainer/policy/normal/std Std 0.552448 +trainer/policy/normal/std Max 6.69507 +trainer/policy/normal/std Min 0.529224 +trainer/policy/normal/log_std Mean 1.02221 +trainer/policy/normal/log_std Std 0.239857 +trainer/policy/normal/log_std Max 1.90137 +trainer/policy/normal/log_std Min -0.636344 +eval/num steps total 115778 +eval/num paths total 116 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.361955 +eval/Actions Std 0.759445 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63426 +time/logging (s) 0.00393753 +time/sampling batch (s) 0.279127 +time/saving (s) 0.00389178 +time/training (s) 7.95059 +time/epoch (s) 10.8718 +time/total (s) 1239.69 +Epoch -885 +---------------------------------- --------------- +2022-05-10 13:31:26.557911 PDT | [1] Epoch -884 finished +---------------------------------- --------------- +epoch -884 +replay_buffer/size 999033 +trainer/num train calls 117000 +trainer/Policy Loss -19.9713 +trainer/Log Pis Mean 24.3514 +trainer/Log Pis Std 13.3253 +trainer/Log Pis Max 60.8767 +trainer/Log Pis Min -9.86876 +trainer/policy/mean Mean -0.026127 +trainer/policy/mean Std 0.908177 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.84514 +trainer/policy/normal/std Std 0.581895 +trainer/policy/normal/std Max 5.15974 +trainer/policy/normal/std Min 0.491493 +trainer/policy/normal/log_std Mean 1.01732 +trainer/policy/normal/log_std Std 0.262069 +trainer/policy/normal/log_std Max 1.64089 +trainer/policy/normal/log_std Min -0.710308 +eval/num steps total 116778 +eval/num paths total 117 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0118554 +eval/Actions Std 0.920249 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60353 +time/logging (s) 0.00414079 +time/sampling batch (s) 0.278949 +time/saving (s) 0.00396353 +time/training (s) 8.58231 +time/epoch (s) 11.4729 +time/total (s) 1251.17 +Epoch -884 +---------------------------------- --------------- +2022-05-10 13:31:37.524056 PDT | [1] Epoch -883 finished +---------------------------------- --------------- +epoch -883 +replay_buffer/size 999033 +trainer/num train calls 118000 +trainer/Policy Loss -18.6589 +trainer/Log Pis Mean 25.657 +trainer/Log Pis Std 14.3549 +trainer/Log Pis Max 81.3638 +trainer/Log Pis Min -8.61966 +trainer/policy/mean Mean -0.0311043 +trainer/policy/mean Std 0.90434 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85509 +trainer/policy/normal/std Std 0.579977 +trainer/policy/normal/std Max 5.15716 +trainer/policy/normal/std Min 0.498922 +trainer/policy/normal/log_std Mean 1.02127 +trainer/policy/normal/log_std Std 0.260159 +trainer/policy/normal/log_std Max 1.64039 +trainer/policy/normal/log_std Min -0.695306 +eval/num steps total 117778 +eval/num paths total 118 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00611316 +eval/Actions Std 0.893529 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.25942 +time/logging (s) 0.00382126 +time/sampling batch (s) 0.279164 +time/saving (s) 0.00354711 +time/training (s) 8.3992 +time/epoch (s) 10.9452 +time/total (s) 1262.12 +Epoch -883 +---------------------------------- --------------- +2022-05-10 13:31:46.731218 PDT | [1] Epoch -882 finished +---------------------------------- --------------- +epoch -882 +replay_buffer/size 999033 +trainer/num train calls 119000 +trainer/Policy Loss -20.4094 +trainer/Log Pis Mean 24.8811 +trainer/Log Pis Std 13.5824 +trainer/Log Pis Max 67.6393 +trainer/Log Pis Min -6.11026 +trainer/policy/mean Mean -0.0201536 +trainer/policy/mean Std 0.908924 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83756 +trainer/policy/normal/std Std 0.575649 +trainer/policy/normal/std Max 5.96935 +trainer/policy/normal/std Min 0.563044 +trainer/policy/normal/log_std Mean 1.01543 +trainer/policy/normal/log_std Std 0.257164 +trainer/policy/normal/log_std Max 1.78664 +trainer/policy/normal/log_std Min -0.574398 +eval/num steps total 118778 +eval/num paths total 119 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121347 +eval/Actions Std 0.899978 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49476 +time/logging (s) 0.00378914 +time/sampling batch (s) 0.277159 +time/saving (s) 0.00348939 +time/training (s) 6.40789 +time/epoch (s) 9.18709 +time/total (s) 1271.31 +Epoch -882 +---------------------------------- --------------- +2022-05-10 13:31:57.999501 PDT | [1] Epoch -881 finished +---------------------------------- --------------- +epoch -881 +replay_buffer/size 999033 +trainer/num train calls 120000 +trainer/Policy Loss -19.5103 +trainer/Log Pis Mean 25.5454 +trainer/Log Pis Std 13.1479 +trainer/Log Pis Max 65.0424 +trainer/Log Pis Min -8.17757 +trainer/policy/mean Mean -0.0333802 +trainer/policy/mean Std 0.905447 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.8788 +trainer/policy/normal/std Std 0.554432 +trainer/policy/normal/std Max 4.79202 +trainer/policy/normal/std Min 0.487438 +trainer/policy/normal/log_std Mean 1.03229 +trainer/policy/normal/log_std Std 0.246978 +trainer/policy/normal/log_std Max 1.56695 +trainer/policy/normal/log_std Min -0.718592 +eval/num steps total 119778 +eval/num paths total 120 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0455937 +eval/Actions Std 0.892578 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61312 +time/logging (s) 0.00414345 +time/sampling batch (s) 0.28156 +time/saving (s) 0.00401431 +time/training (s) 8.34547 +time/epoch (s) 11.2483 +time/total (s) 1282.56 +Epoch -881 +---------------------------------- --------------- +2022-05-10 13:32:08.037007 PDT | [1] Epoch -880 finished +---------------------------------- --------------- +epoch -880 +replay_buffer/size 999033 +trainer/num train calls 121000 +trainer/Policy Loss -18.0144 +trainer/Log Pis Mean 24.538 +trainer/Log Pis Std 12.9336 +trainer/Log Pis Max 66.4304 +trainer/Log Pis Min -6.47458 +trainer/policy/mean Mean -0.0297018 +trainer/policy/mean Std 0.898203 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.88754 +trainer/policy/normal/std Std 0.579212 +trainer/policy/normal/std Max 5.77917 +trainer/policy/normal/std Min 0.448663 +trainer/policy/normal/log_std Mean 1.03294 +trainer/policy/normal/log_std Std 0.259181 +trainer/policy/normal/log_std Max 1.75426 +trainer/policy/normal/log_std Min -0.801484 +eval/num steps total 120778 +eval/num paths total 121 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.139718 +eval/Actions Std 0.874585 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54269 +time/logging (s) 0.00380729 +time/sampling batch (s) 0.277234 +time/saving (s) 0.00346223 +time/training (s) 7.18934 +time/epoch (s) 10.0165 +time/total (s) 1292.58 +Epoch -880 +---------------------------------- --------------- +2022-05-10 13:32:18.440648 PDT | [1] Epoch -879 finished +---------------------------------- --------------- +epoch -879 +replay_buffer/size 999033 +trainer/num train calls 122000 +trainer/Policy Loss -19.8716 +trainer/Log Pis Mean 23.7206 +trainer/Log Pis Std 12.4838 +trainer/Log Pis Max 59.6943 +trainer/Log Pis Min -11.5667 +trainer/policy/mean Mean -0.0566205 +trainer/policy/mean Std 0.905046 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82217 +trainer/policy/normal/std Std 0.560855 +trainer/policy/normal/std Max 5.08127 +trainer/policy/normal/std Min 0.532249 +trainer/policy/normal/log_std Mean 1.01106 +trainer/policy/normal/log_std Std 0.252294 +trainer/policy/normal/log_std Max 1.62556 +trainer/policy/normal/log_std Min -0.630644 +eval/num steps total 121778 +eval/num paths total 122 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0249484 +eval/Actions Std 0.973116 +eval/Actions Max 0.999982 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55524 +time/logging (s) 0.00382963 +time/sampling batch (s) 0.315807 +time/saving (s) 0.00358657 +time/training (s) 7.50451 +time/epoch (s) 10.383 +time/total (s) 1302.96 +Epoch -879 +---------------------------------- --------------- +2022-05-10 13:32:29.378952 PDT | [1] Epoch -878 finished +---------------------------------- --------------- +epoch -878 +replay_buffer/size 999033 +trainer/num train calls 123000 +trainer/Policy Loss -19.3288 +trainer/Log Pis Mean 25.1186 +trainer/Log Pis Std 13.1614 +trainer/Log Pis Max 68.2085 +trainer/Log Pis Min -9.91948 +trainer/policy/mean Mean -0.033405 +trainer/policy/mean Std 0.90595 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.88141 +trainer/policy/normal/std Std 0.603836 +trainer/policy/normal/std Max 4.88295 +trainer/policy/normal/std Min 0.512593 +trainer/policy/normal/log_std Mean 1.02816 +trainer/policy/normal/log_std Std 0.271108 +trainer/policy/normal/log_std Max 1.58575 +trainer/policy/normal/log_std Min -0.668274 +eval/num steps total 122778 +eval/num paths total 123 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.292561 +eval/Actions Std 0.860587 +eval/Actions Max 0.999999 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65576 +time/logging (s) 0.00382571 +time/sampling batch (s) 0.332474 +time/saving (s) 0.00367606 +time/training (s) 7.92165 +time/epoch (s) 10.9174 +time/total (s) 1313.89 +Epoch -878 +---------------------------------- --------------- +2022-05-10 13:32:40.739128 PDT | [1] Epoch -877 finished +---------------------------------- --------------- +epoch -877 +replay_buffer/size 999033 +trainer/num train calls 124000 +trainer/Policy Loss -19.3042 +trainer/Log Pis Mean 24.2123 +trainer/Log Pis Std 12.703 +trainer/Log Pis Max 66.0781 +trainer/Log Pis Min -7.61112 +trainer/policy/mean Mean -0.0391434 +trainer/policy/mean Std 0.905242 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81174 +trainer/policy/normal/std Std 0.569164 +trainer/policy/normal/std Max 5.32102 +trainer/policy/normal/std Min 0.48615 +trainer/policy/normal/log_std Mean 1.00597 +trainer/policy/normal/log_std Std 0.260372 +trainer/policy/normal/log_std Max 1.67167 +trainer/policy/normal/log_std Min -0.721238 +eval/num steps total 123778 +eval/num paths total 124 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.09249 +eval/Actions Std 0.90835 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74913 +time/logging (s) 0.00419199 +time/sampling batch (s) 0.330319 +time/saving (s) 0.00409196 +time/training (s) 8.25188 +time/epoch (s) 11.3396 +time/total (s) 1325.23 +Epoch -877 +---------------------------------- --------------- +2022-05-10 13:32:51.695873 PDT | [1] Epoch -876 finished +---------------------------------- --------------- +epoch -876 +replay_buffer/size 999033 +trainer/num train calls 125000 +trainer/Policy Loss -18.4804 +trainer/Log Pis Mean 24.849 +trainer/Log Pis Std 12.7978 +trainer/Log Pis Max 64.236 +trainer/Log Pis Min -4.68172 +trainer/policy/mean Mean -0.0414118 +trainer/policy/mean Std 0.905969 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.89848 +trainer/policy/normal/std Std 0.552765 +trainer/policy/normal/std Max 4.96806 +trainer/policy/normal/std Min 0.482254 +trainer/policy/normal/log_std Mean 1.03944 +trainer/policy/normal/log_std Std 0.246064 +trainer/policy/normal/log_std Max 1.60303 +trainer/policy/normal/log_std Min -0.729285 +eval/num steps total 124778 +eval/num paths total 125 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.086921 +eval/Actions Std 0.879984 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7232 +time/logging (s) 0.00395134 +time/sampling batch (s) 0.58276 +time/saving (s) 0.00393408 +time/training (s) 7.62102 +time/epoch (s) 10.9349 +time/total (s) 1336.17 +Epoch -876 +---------------------------------- --------------- +2022-05-10 13:33:01.565875 PDT | [1] Epoch -875 finished +---------------------------------- --------------- +epoch -875 +replay_buffer/size 999033 +trainer/num train calls 126000 +trainer/Policy Loss -18.0933 +trainer/Log Pis Mean 25.1085 +trainer/Log Pis Std 12.9561 +trainer/Log Pis Max 69.7007 +trainer/Log Pis Min -7.20277 +trainer/policy/mean Mean -0.0396546 +trainer/policy/mean Std 0.903759 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.92436 +trainer/policy/normal/std Std 0.584706 +trainer/policy/normal/std Max 5.32554 +trainer/policy/normal/std Min 0.457678 +trainer/policy/normal/log_std Mean 1.04548 +trainer/policy/normal/log_std Std 0.26065 +trainer/policy/normal/log_std Max 1.67251 +trainer/policy/normal/log_std Min -0.781589 +eval/num steps total 125778 +eval/num paths total 126 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0768891 +eval/Actions Std 0.915567 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7334 +time/logging (s) 0.0036699 +time/sampling batch (s) 0.294434 +time/saving (s) 0.00342815 +time/training (s) 6.81423 +time/epoch (s) 9.84916 +time/total (s) 1346.02 +Epoch -875 +---------------------------------- --------------- +2022-05-10 13:33:12.730961 PDT | [1] Epoch -874 finished +---------------------------------- --------------- +epoch -874 +replay_buffer/size 999033 +trainer/num train calls 127000 +trainer/Policy Loss -18.2929 +trainer/Log Pis Mean 24.4197 +trainer/Log Pis Std 12.5221 +trainer/Log Pis Max 61.3917 +trainer/Log Pis Min -9.35201 +trainer/policy/mean Mean -0.0296861 +trainer/policy/mean Std 0.899375 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84014 +trainer/policy/normal/std Std 0.579143 +trainer/policy/normal/std Max 6.7033 +trainer/policy/normal/std Min 0.398963 +trainer/policy/normal/log_std Mean 1.01452 +trainer/policy/normal/log_std Std 0.270659 +trainer/policy/normal/log_std Max 1.9026 +trainer/policy/normal/log_std Min -0.918886 +eval/num steps total 126778 +eval/num paths total 127 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0763157 +eval/Actions Std 0.914696 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60672 +time/logging (s) 0.00375573 +time/sampling batch (s) 0.529971 +time/saving (s) 0.00344644 +time/training (s) 8.00105 +time/epoch (s) 11.145 +time/total (s) 1357.17 +Epoch -874 +---------------------------------- --------------- +2022-05-10 13:33:23.711090 PDT | [1] Epoch -873 finished +---------------------------------- --------------- +epoch -873 +replay_buffer/size 999033 +trainer/num train calls 128000 +trainer/Policy Loss -19.3138 +trainer/Log Pis Mean 23.7056 +trainer/Log Pis Std 12.7474 +trainer/Log Pis Max 75.0955 +trainer/Log Pis Min -10.2008 +trainer/policy/mean Mean -0.0269262 +trainer/policy/mean Std 0.899638 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.88894 +trainer/policy/normal/std Std 0.567458 +trainer/policy/normal/std Max 5.47171 +trainer/policy/normal/std Min 0.447629 +trainer/policy/normal/log_std Mean 1.03454 +trainer/policy/normal/log_std Std 0.254488 +trainer/policy/normal/log_std Max 1.69959 +trainer/policy/normal/log_std Min -0.803789 +eval/num steps total 127778 +eval/num paths total 128 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109589 +eval/Actions Std 0.847776 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57719 +time/logging (s) 0.00381003 +time/sampling batch (s) 1.03108 +time/saving (s) 0.00346064 +time/training (s) 7.34455 +time/epoch (s) 10.9601 +time/total (s) 1368.13 +Epoch -873 +---------------------------------- --------------- +2022-05-10 13:33:35.003838 PDT | [1] Epoch -872 finished +---------------------------------- --------------- +epoch -872 +replay_buffer/size 999033 +trainer/num train calls 129000 +trainer/Policy Loss -18.7771 +trainer/Log Pis Mean 24.3048 +trainer/Log Pis Std 12.8372 +trainer/Log Pis Max 66.7507 +trainer/Log Pis Min -5.9532 +trainer/policy/mean Mean -0.0556862 +trainer/policy/mean Std 0.90295 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89386 +trainer/policy/normal/std Std 0.589685 +trainer/policy/normal/std Max 4.98013 +trainer/policy/normal/std Min 0.408698 +trainer/policy/normal/log_std Mean 1.03283 +trainer/policy/normal/log_std Std 0.273809 +trainer/policy/normal/log_std Max 1.60546 +trainer/policy/normal/log_std Min -0.894778 +eval/num steps total 128778 +eval/num paths total 129 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0886179 +eval/Actions Std 0.910309 +eval/Actions Max 0.99999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5011 +time/logging (s) 0.00411737 +time/sampling batch (s) 1.03287 +time/saving (s) 0.00383565 +time/training (s) 7.73078 +time/epoch (s) 11.2727 +time/total (s) 1379.41 +Epoch -872 +---------------------------------- --------------- +2022-05-10 13:33:44.820336 PDT | [1] Epoch -871 finished +---------------------------------- --------------- +epoch -871 +replay_buffer/size 999033 +trainer/num train calls 130000 +trainer/Policy Loss -19.5852 +trainer/Log Pis Mean 24.1903 +trainer/Log Pis Std 13.5533 +trainer/Log Pis Max 69.1383 +trainer/Log Pis Min -6.5674 +trainer/policy/mean Mean -0.0361984 +trainer/policy/mean Std 0.905396 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86639 +trainer/policy/normal/std Std 0.561591 +trainer/policy/normal/std Max 5.55715 +trainer/policy/normal/std Min 0.477879 +trainer/policy/normal/log_std Mean 1.02704 +trainer/policy/normal/log_std Std 0.251779 +trainer/policy/normal/log_std Max 1.71509 +trainer/policy/normal/log_std Min -0.738399 +eval/num steps total 129778 +eval/num paths total 130 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0438577 +eval/Actions Std 0.89488 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83758 +time/logging (s) 0.00381789 +time/sampling batch (s) 0.279153 +time/saving (s) 0.00376676 +time/training (s) 6.67131 +time/epoch (s) 9.79562 +time/total (s) 1389.2 +Epoch -871 +---------------------------------- --------------- +2022-05-10 13:33:55.165363 PDT | [1] Epoch -870 finished +---------------------------------- --------------- +epoch -870 +replay_buffer/size 999033 +trainer/num train calls 131000 +trainer/Policy Loss -18.5727 +trainer/Log Pis Mean 25.0217 +trainer/Log Pis Std 13.2144 +trainer/Log Pis Max 76.6792 +trainer/Log Pis Min -10.3398 +trainer/policy/mean Mean -0.0456766 +trainer/policy/mean Std 0.907376 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8533 +trainer/policy/normal/std Std 0.569971 +trainer/policy/normal/std Max 4.9663 +trainer/policy/normal/std Min 0.387815 +trainer/policy/normal/log_std Mean 1.02104 +trainer/policy/normal/log_std Std 0.258988 +trainer/policy/normal/log_std Max 1.60267 +trainer/policy/normal/log_std Min -0.947226 +eval/num steps total 130778 +eval/num paths total 131 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11792 +eval/Actions Std 0.907563 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48296 +time/logging (s) 0.00378162 +time/sampling batch (s) 0.279036 +time/saving (s) 0.00343899 +time/training (s) 7.55547 +time/epoch (s) 10.3247 +time/total (s) 1399.53 +Epoch -870 +---------------------------------- --------------- +2022-05-10 13:34:06.306763 PDT | [1] Epoch -869 finished +---------------------------------- --------------- +epoch -869 +replay_buffer/size 999033 +trainer/num train calls 132000 +trainer/Policy Loss -18.5663 +trainer/Log Pis Mean 23.7108 +trainer/Log Pis Std 13.4369 +trainer/Log Pis Max 68.2554 +trainer/Log Pis Min -11.6165 +trainer/policy/mean Mean -0.0324133 +trainer/policy/mean Std 0.900309 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87477 +trainer/policy/normal/std Std 0.554745 +trainer/policy/normal/std Max 5.57858 +trainer/policy/normal/std Min 0.465614 +trainer/policy/normal/log_std Mean 1.03095 +trainer/policy/normal/log_std Std 0.246107 +trainer/policy/normal/log_std Max 1.71893 +trainer/policy/normal/log_std Min -0.764398 +eval/num steps total 131778 +eval/num paths total 132 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0731495 +eval/Actions Std 0.902305 +eval/Actions Max 0.999981 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6622 +time/logging (s) 0.00367845 +time/sampling batch (s) 0.28437 +time/saving (s) 0.00340888 +time/training (s) 8.16722 +time/epoch (s) 11.1209 +time/total (s) 1410.66 +Epoch -869 +---------------------------------- --------------- +2022-05-10 13:34:17.172179 PDT | [1] Epoch -868 finished +---------------------------------- --------------- +epoch -868 +replay_buffer/size 999033 +trainer/num train calls 133000 +trainer/Policy Loss -19.2921 +trainer/Log Pis Mean 24.1546 +trainer/Log Pis Std 12.3283 +trainer/Log Pis Max 68.1101 +trainer/Log Pis Min -6.94846 +trainer/policy/mean Mean -0.0223337 +trainer/policy/mean Std 0.905935 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86213 +trainer/policy/normal/std Std 0.571612 +trainer/policy/normal/std Max 7.27479 +trainer/policy/normal/std Min 0.387423 +trainer/policy/normal/log_std Mean 1.02538 +trainer/policy/normal/log_std Std 0.250746 +trainer/policy/normal/log_std Max 1.98442 +trainer/policy/normal/log_std Min -0.948239 +eval/num steps total 132778 +eval/num paths total 133 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.319197 +eval/Actions Std 0.87298 +eval/Actions Max 0.999966 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60794 +time/logging (s) 0.00374818 +time/sampling batch (s) 0.78077 +time/saving (s) 0.00345468 +time/training (s) 7.44934 +time/epoch (s) 10.8452 +time/total (s) 1421.5 +Epoch -868 +---------------------------------- --------------- +2022-05-10 13:34:28.336969 PDT | [1] Epoch -867 finished +---------------------------------- --------------- +epoch -867 +replay_buffer/size 999033 +trainer/num train calls 134000 +trainer/Policy Loss -19.8525 +trainer/Log Pis Mean 24.961 +trainer/Log Pis Std 12.3823 +trainer/Log Pis Max 73.1064 +trainer/Log Pis Min -10.0975 +trainer/policy/mean Mean -0.0345498 +trainer/policy/mean Std 0.902481 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84923 +trainer/policy/normal/std Std 0.571089 +trainer/policy/normal/std Max 5.36341 +trainer/policy/normal/std Min 0.486125 +trainer/policy/normal/log_std Mean 1.01959 +trainer/policy/normal/log_std Std 0.259149 +trainer/policy/normal/log_std Max 1.6796 +trainer/policy/normal/log_std Min -0.721289 +eval/num steps total 133778 +eval/num paths total 134 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.44659 +eval/Actions Std 0.877091 +eval/Actions Max 0.999989 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69144 +time/logging (s) 0.00369076 +time/sampling batch (s) 0.530613 +time/saving (s) 0.00346871 +time/training (s) 7.91475 +time/epoch (s) 11.144 +time/total (s) 1432.65 +Epoch -867 +---------------------------------- --------------- +2022-05-10 13:34:40.006007 PDT | [1] Epoch -866 finished +---------------------------------- --------------- +epoch -866 +replay_buffer/size 999033 +trainer/num train calls 135000 +trainer/Policy Loss -19.2631 +trainer/Log Pis Mean 24.3696 +trainer/Log Pis Std 13.7352 +trainer/Log Pis Max 69.1637 +trainer/Log Pis Min -13.9757 +trainer/policy/mean Mean -0.0312076 +trainer/policy/mean Std 0.905778 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8204 +trainer/policy/normal/std Std 0.586327 +trainer/policy/normal/std Max 5.42957 +trainer/policy/normal/std Min 0.419419 +trainer/policy/normal/log_std Mean 1.00617 +trainer/policy/normal/log_std Std 0.277172 +trainer/policy/normal/log_std Max 1.69186 +trainer/policy/normal/log_std Min -0.868886 +eval/num steps total 134778 +eval/num paths total 135 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0590053 +eval/Actions Std 0.905505 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71884 +time/logging (s) 0.00417211 +time/sampling batch (s) 1.33382 +time/saving (s) 0.00424839 +time/training (s) 7.58743 +time/epoch (s) 11.6485 +time/total (s) 1444.3 +Epoch -866 +---------------------------------- --------------- +2022-05-10 13:34:50.866116 PDT | [1] Epoch -865 finished +---------------------------------- --------------- +epoch -865 +replay_buffer/size 999033 +trainer/num train calls 136000 +trainer/Policy Loss -19.6246 +trainer/Log Pis Mean 24.0864 +trainer/Log Pis Std 12.9143 +trainer/Log Pis Max 65.773 +trainer/Log Pis Min -7.55613 +trainer/policy/mean Mean -0.0521924 +trainer/policy/mean Std 0.903597 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.84533 +trainer/policy/normal/std Std 0.576851 +trainer/policy/normal/std Max 6.1712 +trainer/policy/normal/std Min 0.403603 +trainer/policy/normal/log_std Mean 1.0175 +trainer/policy/normal/log_std Std 0.263735 +trainer/policy/normal/log_std Max 1.81989 +trainer/policy/normal/log_std Min -0.907324 +eval/num steps total 135778 +eval/num paths total 136 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0769626 +eval/Actions Std 0.897511 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88188 +time/logging (s) 0.00378286 +time/sampling batch (s) 0.331655 +time/saving (s) 0.00371811 +time/training (s) 7.61727 +time/epoch (s) 10.8383 +time/total (s) 1455.14 +Epoch -865 +---------------------------------- --------------- +2022-05-10 13:35:01.074781 PDT | [1] Epoch -864 finished +---------------------------------- --------------- +epoch -864 +replay_buffer/size 999033 +trainer/num train calls 137000 +trainer/Policy Loss -18.4367 +trainer/Log Pis Mean 22.8031 +trainer/Log Pis Std 13.3466 +trainer/Log Pis Max 75.0301 +trainer/Log Pis Min -15.5101 +trainer/policy/mean Mean -0.0212501 +trainer/policy/mean Std 0.902379 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82667 +trainer/policy/normal/std Std 0.546751 +trainer/policy/normal/std Max 5.08107 +trainer/policy/normal/std Min 0.486756 +trainer/policy/normal/log_std Mean 1.01427 +trainer/policy/normal/log_std Std 0.244787 +trainer/policy/normal/log_std Max 1.62552 +trainer/policy/normal/log_std Min -0.719992 +eval/num steps total 136778 +eval/num paths total 137 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0961423 +eval/Actions Std 0.889074 +eval/Actions Max 0.999995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73294 +time/logging (s) 0.00380455 +time/sampling batch (s) 0.333399 +time/saving (s) 0.00371567 +time/training (s) 7.11358 +time/epoch (s) 10.1874 +time/total (s) 1465.33 +Epoch -864 +---------------------------------- --------------- +2022-05-10 13:35:12.340141 PDT | [1] Epoch -863 finished +---------------------------------- --------------- +epoch -863 +replay_buffer/size 999033 +trainer/num train calls 138000 +trainer/Policy Loss -18.9305 +trainer/Log Pis Mean 24.3588 +trainer/Log Pis Std 13.9746 +trainer/Log Pis Max 72.6625 +trainer/Log Pis Min -5.89894 +trainer/policy/mean Mean -0.062024 +trainer/policy/mean Std 0.902038 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89338 +trainer/policy/normal/std Std 0.587928 +trainer/policy/normal/std Max 5.50443 +trainer/policy/normal/std Min 0.371083 +trainer/policy/normal/log_std Mean 1.03409 +trainer/policy/normal/log_std Std 0.264102 +trainer/policy/normal/log_std Max 1.70555 +trainer/policy/normal/log_std Min -0.991328 +eval/num steps total 137778 +eval/num paths total 138 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.205836 +eval/Actions Std 0.904252 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.98534 +time/logging (s) 0.00380875 +time/sampling batch (s) 0.336652 +time/saving (s) 0.003666 +time/training (s) 7.91459 +time/epoch (s) 11.2441 +time/total (s) 1476.58 +Epoch -863 +---------------------------------- --------------- +2022-05-10 13:35:23.837596 PDT | [1] Epoch -862 finished +---------------------------------- --------------- +epoch -862 +replay_buffer/size 999033 +trainer/num train calls 139000 +trainer/Policy Loss -19.4455 +trainer/Log Pis Mean 24.162 +trainer/Log Pis Std 13.1435 +trainer/Log Pis Max 76.9147 +trainer/Log Pis Min -6.18063 +trainer/policy/mean Mean -0.0281489 +trainer/policy/mean Std 0.90313 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.89208 +trainer/policy/normal/std Std 0.57739 +trainer/policy/normal/std Max 5.12387 +trainer/policy/normal/std Min 0.445786 +trainer/policy/normal/log_std Mean 1.03374 +trainer/policy/normal/log_std Std 0.265835 +trainer/policy/normal/log_std Max 1.63391 +trainer/policy/normal/log_std Min -0.807916 +eval/num steps total 138778 +eval/num paths total 139 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128712 +eval/Actions Std 0.891813 +eval/Actions Max 0.99997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8592 +time/logging (s) 0.00413096 +time/sampling batch (s) 0.583205 +time/saving (s) 0.00405276 +time/training (s) 8.02591 +time/epoch (s) 11.4765 +time/total (s) 1488.06 +Epoch -862 +---------------------------------- --------------- +2022-05-10 13:35:34.607593 PDT | [1] Epoch -861 finished +---------------------------------- --------------- +epoch -861 +replay_buffer/size 999033 +trainer/num train calls 140000 +trainer/Policy Loss -19.8081 +trainer/Log Pis Mean 25.122 +trainer/Log Pis Std 13.0786 +trainer/Log Pis Max 74.3625 +trainer/Log Pis Min -5.70261 +trainer/policy/mean Mean -0.0450587 +trainer/policy/mean Std 0.91015 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86273 +trainer/policy/normal/std Std 0.557407 +trainer/policy/normal/std Max 4.8963 +trainer/policy/normal/std Min 0.496669 +trainer/policy/normal/log_std Mean 1.02516 +trainer/policy/normal/log_std Std 0.256693 +trainer/policy/normal/log_std Max 1.58848 +trainer/policy/normal/log_std Min -0.699832 +eval/num steps total 139778 +eval/num paths total 140 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.26428 +eval/Actions Std 0.836762 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61184 +time/logging (s) 0.00460649 +time/sampling batch (s) 0.336886 +time/saving (s) 0.00392993 +time/training (s) 7.79162 +time/epoch (s) 10.7489 +time/total (s) 1498.81 +Epoch -861 +---------------------------------- --------------- +2022-05-10 13:35:46.150790 PDT | [1] Epoch -860 finished +---------------------------------- --------------- +epoch -860 +replay_buffer/size 999033 +trainer/num train calls 141000 +trainer/Policy Loss -19.1225 +trainer/Log Pis Mean 25.0029 +trainer/Log Pis Std 13.3003 +trainer/Log Pis Max 62.9405 +trainer/Log Pis Min -4.98961 +trainer/policy/mean Mean -0.0586805 +trainer/policy/mean Std 0.903033 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85627 +trainer/policy/normal/std Std 0.57067 +trainer/policy/normal/std Max 4.83417 +trainer/policy/normal/std Min 0.48214 +trainer/policy/normal/log_std Mean 1.02159 +trainer/policy/normal/log_std Std 0.262997 +trainer/policy/normal/log_std Max 1.57571 +trainer/policy/normal/log_std Min -0.72952 +eval/num steps total 140778 +eval/num paths total 141 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.103061 +eval/Actions Std 0.862978 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84542 +time/logging (s) 0.00368489 +time/sampling batch (s) 0.556294 +time/saving (s) 0.00338821 +time/training (s) 8.11266 +time/epoch (s) 11.5215 +time/total (s) 1510.34 +Epoch -860 +---------------------------------- --------------- +2022-05-10 13:35:56.624145 PDT | [1] Epoch -859 finished +---------------------------------- --------------- +epoch -859 +replay_buffer/size 999033 +trainer/num train calls 142000 +trainer/Policy Loss -19.4713 +trainer/Log Pis Mean 22.9947 +trainer/Log Pis Std 13.3058 +trainer/Log Pis Max 62.0376 +trainer/Log Pis Min -6.92078 +trainer/policy/mean Mean -0.0355645 +trainer/policy/mean Std 0.899809 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.8384 +trainer/policy/normal/std Std 0.603078 +trainer/policy/normal/std Max 5.09421 +trainer/policy/normal/std Min 0.360223 +trainer/policy/normal/log_std Mean 1.01108 +trainer/policy/normal/log_std Std 0.283912 +trainer/policy/normal/log_std Max 1.6281 +trainer/policy/normal/log_std Min -1.02103 +eval/num steps total 141778 +eval/num paths total 142 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.281872 +eval/Actions Std 0.885722 +eval/Actions Max 0.99999 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49858 +time/logging (s) 0.00369399 +time/sampling batch (s) 0.77901 +time/saving (s) 0.00343073 +time/training (s) 7.16844 +time/epoch (s) 10.4532 +time/total (s) 1520.79 +Epoch -859 +---------------------------------- --------------- +2022-05-10 13:36:06.834930 PDT | [1] Epoch -858 finished +---------------------------------- --------------- +epoch -858 +replay_buffer/size 999033 +trainer/num train calls 143000 +trainer/Policy Loss -19.8364 +trainer/Log Pis Mean 24.4131 +trainer/Log Pis Std 13.2955 +trainer/Log Pis Max 74.5923 +trainer/Log Pis Min -8.76992 +trainer/policy/mean Mean -0.0482976 +trainer/policy/mean Std 0.908382 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84572 +trainer/policy/normal/std Std 0.58363 +trainer/policy/normal/std Max 6.02109 +trainer/policy/normal/std Min 0.396367 +trainer/policy/normal/log_std Mean 1.01675 +trainer/policy/normal/log_std Std 0.268256 +trainer/policy/normal/log_std Max 1.79527 +trainer/policy/normal/log_std Min -0.925414 +eval/num steps total 142778 +eval/num paths total 143 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.267728 +eval/Actions Std 0.873098 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70085 +time/logging (s) 0.00381483 +time/sampling batch (s) 0.777765 +time/saving (s) 0.0034358 +time/training (s) 6.70489 +time/epoch (s) 10.1908 +time/total (s) 1530.99 +Epoch -858 +---------------------------------- --------------- +2022-05-10 13:36:16.895234 PDT | [1] Epoch -857 finished +---------------------------------- -------------- +epoch -857 +replay_buffer/size 999033 +trainer/num train calls 144000 +trainer/Policy Loss -19.4425 +trainer/Log Pis Mean 24.0888 +trainer/Log Pis Std 13.424 +trainer/Log Pis Max 65.8917 +trainer/Log Pis Min -7.44767 +trainer/policy/mean Mean -0.060561 +trainer/policy/mean Std 0.90439 +trainer/policy/mean Max 0.999969 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83902 +trainer/policy/normal/std Std 0.589854 +trainer/policy/normal/std Max 5.58293 +trainer/policy/normal/std Min 0.441284 +trainer/policy/normal/log_std Mean 1.01293 +trainer/policy/normal/log_std Std 0.276246 +trainer/policy/normal/log_std Max 1.71971 +trainer/policy/normal/log_std Min -0.818068 +eval/num steps total 143778 +eval/num paths total 144 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.13993 +eval/Actions Std 0.891107 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58435 +time/logging (s) 0.0037408 +time/sampling batch (s) 0.52701 +time/saving (s) 0.0034617 +time/training (s) 6.9216 +time/epoch (s) 10.0402 +time/total (s) 1541.03 +Epoch -857 +---------------------------------- -------------- +2022-05-10 13:36:26.719422 PDT | [1] Epoch -856 finished +---------------------------------- --------------- +epoch -856 +replay_buffer/size 999033 +trainer/num train calls 145000 +trainer/Policy Loss -19.5099 +trainer/Log Pis Mean 24.1222 +trainer/Log Pis Std 12.6703 +trainer/Log Pis Max 66.5024 +trainer/Log Pis Min -6.48046 +trainer/policy/mean Mean -0.0552328 +trainer/policy/mean Std 0.902528 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87736 +trainer/policy/normal/std Std 0.607779 +trainer/policy/normal/std Max 5.89962 +trainer/policy/normal/std Min 0.473817 +trainer/policy/normal/log_std Mean 1.02612 +trainer/policy/normal/log_std Std 0.274969 +trainer/policy/normal/log_std Max 1.77489 +trainer/policy/normal/log_std Min -0.746935 +eval/num steps total 144778 +eval/num paths total 145 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.446588 +eval/Actions Std 0.801517 +eval/Actions Max 0.999998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52392 +time/logging (s) 0.00418071 +time/sampling batch (s) 0.533798 +time/saving (s) 0.00423411 +time/training (s) 6.73795 +time/epoch (s) 9.80408 +time/total (s) 1550.84 +Epoch -856 +---------------------------------- --------------- +2022-05-10 13:36:36.425235 PDT | [1] Epoch -855 finished +---------------------------------- --------------- +epoch -855 +replay_buffer/size 999033 +trainer/num train calls 146000 +trainer/Policy Loss -20.5812 +trainer/Log Pis Mean 25.3912 +trainer/Log Pis Std 14.0038 +trainer/Log Pis Max 88.3302 +trainer/Log Pis Min -9.38697 +trainer/policy/mean Mean -0.0421874 +trainer/policy/mean Std 0.906699 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.86919 +trainer/policy/normal/std Std 0.585787 +trainer/policy/normal/std Max 5.25736 +trainer/policy/normal/std Min 0.477678 +trainer/policy/normal/log_std Mean 1.02502 +trainer/policy/normal/log_std Std 0.267557 +trainer/policy/normal/log_std Max 1.65963 +trainer/policy/normal/log_std Min -0.738818 +eval/num steps total 145778 +eval/num paths total 146 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.265862 +eval/Actions Std 0.870818 +eval/Actions Max 0.999974 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60971 +time/logging (s) 0.00409365 +time/sampling batch (s) 0.283246 +time/saving (s) 0.0037174 +time/training (s) 6.78404 +time/epoch (s) 9.68481 +time/total (s) 1560.53 +Epoch -855 +---------------------------------- --------------- +2022-05-10 13:36:45.351928 PDT | [1] Epoch -854 finished +---------------------------------- --------------- +epoch -854 +replay_buffer/size 999033 +trainer/num train calls 147000 +trainer/Policy Loss -19.2956 +trainer/Log Pis Mean 24.7052 +trainer/Log Pis Std 12.9869 +trainer/Log Pis Max 63.7081 +trainer/Log Pis Min -4.13068 +trainer/policy/mean Mean -0.03257 +trainer/policy/mean Std 0.905326 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85885 +trainer/policy/normal/std Std 0.581875 +trainer/policy/normal/std Max 5.40411 +trainer/policy/normal/std Min 0.318735 +trainer/policy/normal/log_std Mean 1.02219 +trainer/policy/normal/log_std Std 0.263373 +trainer/policy/normal/log_std Max 1.68716 +trainer/policy/normal/log_std Min -1.14339 +eval/num steps total 146778 +eval/num paths total 147 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.256289 +eval/Actions Std 0.876368 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31124 +time/logging (s) 0.00372599 +time/sampling batch (s) 0.282292 +time/saving (s) 0.00348303 +time/training (s) 6.30457 +time/epoch (s) 8.90531 +time/total (s) 1569.44 +Epoch -854 +---------------------------------- --------------- +2022-05-10 13:36:56.074818 PDT | [1] Epoch -853 finished +---------------------------------- --------------- +epoch -853 +replay_buffer/size 999033 +trainer/num train calls 148000 +trainer/Policy Loss -18.8313 +trainer/Log Pis Mean 23.4288 +trainer/Log Pis Std 12.5041 +trainer/Log Pis Max 66.1386 +trainer/Log Pis Min -5.22156 +trainer/policy/mean Mean -0.0391306 +trainer/policy/mean Std 0.904394 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83933 +trainer/policy/normal/std Std 0.589405 +trainer/policy/normal/std Max 5.66827 +trainer/policy/normal/std Min 0.401809 +trainer/policy/normal/log_std Mean 1.01347 +trainer/policy/normal/log_std Std 0.273454 +trainer/policy/normal/log_std Max 1.73488 +trainer/policy/normal/log_std Min -0.911778 +eval/num steps total 147778 +eval/num paths total 148 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133134 +eval/Actions Std 0.904336 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48448 +time/logging (s) 0.00383398 +time/sampling batch (s) 0.800669 +time/saving (s) 0.00369075 +time/training (s) 7.40906 +time/epoch (s) 10.7017 +time/total (s) 1580.14 +Epoch -853 +---------------------------------- --------------- +2022-05-10 13:37:07.521578 PDT | [1] Epoch -852 finished +---------------------------------- --------------- +epoch -852 +replay_buffer/size 999033 +trainer/num train calls 149000 +trainer/Policy Loss -19.4174 +trainer/Log Pis Mean 24.6124 +trainer/Log Pis Std 12.666 +trainer/Log Pis Max 64.9857 +trainer/Log Pis Min -9.01013 +trainer/policy/mean Mean -0.0330442 +trainer/policy/mean Std 0.900169 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86543 +trainer/policy/normal/std Std 0.581317 +trainer/policy/normal/std Max 5.38891 +trainer/policy/normal/std Min 0.419732 +trainer/policy/normal/log_std Mean 1.02452 +trainer/policy/normal/log_std Std 0.263781 +trainer/policy/normal/log_std Max 1.68434 +trainer/policy/normal/log_std Min -0.868139 +eval/num steps total 148778 +eval/num paths total 149 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.327688 +eval/Actions Std 0.911908 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63166 +time/logging (s) 0.00384689 +time/sampling batch (s) 0.586685 +time/saving (s) 0.00375823 +time/training (s) 8.19959 +time/epoch (s) 11.4255 +time/total (s) 1591.57 +Epoch -852 +---------------------------------- --------------- +2022-05-10 13:37:18.969004 PDT | [1] Epoch -851 finished +---------------------------------- --------------- +epoch -851 +replay_buffer/size 999033 +trainer/num train calls 150000 +trainer/Policy Loss -19.352 +trainer/Log Pis Mean 24.5297 +trainer/Log Pis Std 13.5349 +trainer/Log Pis Max 67.8813 +trainer/Log Pis Min -8.17066 +trainer/policy/mean Mean -0.0556521 +trainer/policy/mean Std 0.905271 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.91986 +trainer/policy/normal/std Std 0.590666 +trainer/policy/normal/std Max 5.83797 +trainer/policy/normal/std Min 0.434913 +trainer/policy/normal/log_std Mean 1.04258 +trainer/policy/normal/log_std Std 0.269005 +trainer/policy/normal/log_std Max 1.76438 +trainer/policy/normal/log_std Min -0.832608 +eval/num steps total 149778 +eval/num paths total 150 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0575401 +eval/Actions Std 0.911553 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.9037 +time/logging (s) 0.00400448 +time/sampling batch (s) 0.295442 +time/saving (s) 0.00386065 +time/training (s) 8.21939 +time/epoch (s) 11.4264 +time/total (s) 1603 +Epoch -851 +---------------------------------- --------------- +2022-05-10 13:37:29.771560 PDT | [1] Epoch -850 finished +---------------------------------- --------------- +epoch -850 +replay_buffer/size 999033 +trainer/num train calls 151000 +trainer/Policy Loss -20.2225 +trainer/Log Pis Mean 24.321 +trainer/Log Pis Std 12.7945 +trainer/Log Pis Max 66.2935 +trainer/Log Pis Min -6.87912 +trainer/policy/mean Mean -0.051601 +trainer/policy/mean Std 0.907037 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88647 +trainer/policy/normal/std Std 0.59231 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.429096 +trainer/policy/normal/log_std Mean 1.03091 +trainer/policy/normal/log_std Std 0.268348 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.846074 +eval/num steps total 150778 +eval/num paths total 151 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.234386 +eval/Actions Std 0.872547 +eval/Actions Max 0.999988 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72324 +time/logging (s) 0.00373854 +time/sampling batch (s) 0.526466 +time/saving (s) 0.00357633 +time/training (s) 7.52472 +time/epoch (s) 10.7817 +time/total (s) 1613.78 +Epoch -850 +---------------------------------- --------------- +2022-05-10 13:37:39.969809 PDT | [1] Epoch -849 finished +---------------------------------- -------------- +epoch -849 +replay_buffer/size 999033 +trainer/num train calls 152000 +trainer/Policy Loss -20.2512 +trainer/Log Pis Mean 23.9375 +trainer/Log Pis Std 12.631 +trainer/Log Pis Max 65.7839 +trainer/Log Pis Min -7.79444 +trainer/policy/mean Mean -0.0320063 +trainer/policy/mean Std 0.906541 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8448 +trainer/policy/normal/std Std 0.556969 +trainer/policy/normal/std Max 5.62536 +trainer/policy/normal/std Min 0.474331 +trainer/policy/normal/log_std Mean 1.01926 +trainer/policy/normal/log_std Std 0.253722 +trainer/policy/normal/log_std Max 1.72729 +trainer/policy/normal/log_std Min -0.745851 +eval/num steps total 151778 +eval/num paths total 152 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0382825 +eval/Actions Std 0.898304 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45943 +time/logging (s) 0.0036677 +time/sampling batch (s) 0.276507 +time/saving (s) 0.0034305 +time/training (s) 7.43493 +time/epoch (s) 10.178 +time/total (s) 1623.96 +Epoch -849 +---------------------------------- -------------- +2022-05-10 13:37:50.595722 PDT | [1] Epoch -848 finished +---------------------------------- --------------- +epoch -848 +replay_buffer/size 999033 +trainer/num train calls 153000 +trainer/Policy Loss -20.4046 +trainer/Log Pis Mean 24.2015 +trainer/Log Pis Std 12.9033 +trainer/Log Pis Max 74.8875 +trainer/Log Pis Min -7.83828 +trainer/policy/mean Mean -0.0657141 +trainer/policy/mean Std 0.903739 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86337 +trainer/policy/normal/std Std 0.584684 +trainer/policy/normal/std Max 6.52537 +trainer/policy/normal/std Min 0.423845 +trainer/policy/normal/log_std Mean 1.02374 +trainer/policy/normal/log_std Std 0.263389 +trainer/policy/normal/log_std Max 1.8757 +trainer/policy/normal/log_std Min -0.858388 +eval/num steps total 152778 +eval/num paths total 153 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0410385 +eval/Actions Std 0.896194 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51791 +time/logging (s) 0.00394347 +time/sampling batch (s) 0.527638 +time/saving (s) 0.00360795 +time/training (s) 7.5529 +time/epoch (s) 10.606 +time/total (s) 1634.57 +Epoch -848 +---------------------------------- --------------- +2022-05-10 13:38:00.427053 PDT | [1] Epoch -847 finished +---------------------------------- --------------- +epoch -847 +replay_buffer/size 999033 +trainer/num train calls 154000 +trainer/Policy Loss -19.4675 +trainer/Log Pis Mean 24.3286 +trainer/Log Pis Std 13.4211 +trainer/Log Pis Max 80.0596 +trainer/Log Pis Min -6.10636 +trainer/policy/mean Mean -0.0240179 +trainer/policy/mean Std 0.901477 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82171 +trainer/policy/normal/std Std 0.580457 +trainer/policy/normal/std Max 5.02567 +trainer/policy/normal/std Min 0.402997 +trainer/policy/normal/log_std Mean 1.00744 +trainer/policy/normal/log_std Std 0.27284 +trainer/policy/normal/log_std Max 1.61456 +trainer/policy/normal/log_std Min -0.908826 +eval/num steps total 153778 +eval/num paths total 154 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106084 +eval/Actions Std 0.906663 +eval/Actions Max 0.999981 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37988 +time/logging (s) 0.00425643 +time/sampling batch (s) 0.793341 +time/saving (s) 0.00411311 +time/training (s) 6.62878 +time/epoch (s) 9.81037 +time/total (s) 1644.39 +Epoch -847 +---------------------------------- --------------- +2022-05-10 13:38:11.016184 PDT | [1] Epoch -846 finished +---------------------------------- --------------- +epoch -846 +replay_buffer/size 999033 +trainer/num train calls 155000 +trainer/Policy Loss -19.4417 +trainer/Log Pis Mean 23.4809 +trainer/Log Pis Std 12.5227 +trainer/Log Pis Max 64.3279 +trainer/Log Pis Min -8.84078 +trainer/policy/mean Mean -0.0423805 +trainer/policy/mean Std 0.903786 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85038 +trainer/policy/normal/std Std 0.577219 +trainer/policy/normal/std Max 4.96145 +trainer/policy/normal/std Min 0.487648 +trainer/policy/normal/log_std Mean 1.01814 +trainer/policy/normal/log_std Std 0.270996 +trainer/policy/normal/log_std Max 1.6017 +trainer/policy/normal/log_std Min -0.718161 +eval/num steps total 154778 +eval/num paths total 155 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113196 +eval/Actions Std 0.864911 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73741 +time/logging (s) 0.00426973 +time/sampling batch (s) 0.781649 +time/saving (s) 0.00415352 +time/training (s) 7.04053 +time/epoch (s) 10.568 +time/total (s) 1654.96 +Epoch -846 +---------------------------------- --------------- +2022-05-10 13:38:20.353249 PDT | [1] Epoch -845 finished +---------------------------------- --------------- +epoch -845 +replay_buffer/size 999033 +trainer/num train calls 156000 +trainer/Policy Loss -19.566 +trainer/Log Pis Mean 24.699 +trainer/Log Pis Std 12.6599 +trainer/Log Pis Max 68.0552 +trainer/Log Pis Min -4.72468 +trainer/policy/mean Mean -0.0205801 +trainer/policy/mean Std 0.903234 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85615 +trainer/policy/normal/std Std 0.580535 +trainer/policy/normal/std Max 5.05819 +trainer/policy/normal/std Min 0.467519 +trainer/policy/normal/log_std Mean 1.02128 +trainer/policy/normal/log_std Std 0.262501 +trainer/policy/normal/log_std Max 1.62101 +trainer/policy/normal/log_std Min -0.760315 +eval/num steps total 155778 +eval/num paths total 156 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.14002 +eval/Actions Std 0.883774 +eval/Actions Max 0.999995 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37797 +time/logging (s) 0.00427954 +time/sampling batch (s) 0.279811 +time/saving (s) 0.00399841 +time/training (s) 6.64968 +time/epoch (s) 9.31574 +time/total (s) 1664.28 +Epoch -845 +---------------------------------- --------------- +2022-05-10 13:38:29.197379 PDT | [1] Epoch -844 finished +---------------------------------- --------------- +epoch -844 +replay_buffer/size 999033 +trainer/num train calls 157000 +trainer/Policy Loss -17.9135 +trainer/Log Pis Mean 24.2029 +trainer/Log Pis Std 12.6003 +trainer/Log Pis Max 64.5727 +trainer/Log Pis Min -4.6479 +trainer/policy/mean Mean -0.0645303 +trainer/policy/mean Std 0.898999 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84402 +trainer/policy/normal/std Std 0.600985 +trainer/policy/normal/std Max 5.61312 +trainer/policy/normal/std Min 0.45708 +trainer/policy/normal/log_std Mean 1.01404 +trainer/policy/normal/log_std Std 0.277766 +trainer/policy/normal/log_std Max 1.72511 +trainer/policy/normal/log_std Min -0.782898 +eval/num steps total 156778 +eval/num paths total 157 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.293366 +eval/Actions Std 0.822364 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43497 +time/logging (s) 0.0038036 +time/sampling batch (s) 0.278632 +time/saving (s) 0.00356469 +time/training (s) 6.10184 +time/epoch (s) 8.82281 +time/total (s) 1673.1 +Epoch -844 +---------------------------------- --------------- +2022-05-10 13:38:38.920901 PDT | [1] Epoch -843 finished +---------------------------------- --------------- +epoch -843 +replay_buffer/size 999033 +trainer/num train calls 158000 +trainer/Policy Loss -20.2238 +trainer/Log Pis Mean 24.2644 +trainer/Log Pis Std 13.3985 +trainer/Log Pis Max 73.0604 +trainer/Log Pis Min -7.17759 +trainer/policy/mean Mean -0.0587804 +trainer/policy/mean Std 0.908339 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85947 +trainer/policy/normal/std Std 0.593682 +trainer/policy/normal/std Max 5.2668 +trainer/policy/normal/std Min 0.45894 +trainer/policy/normal/log_std Mean 1.02021 +trainer/policy/normal/log_std Std 0.275906 +trainer/policy/normal/log_std Max 1.66142 +trainer/policy/normal/log_std Min -0.778835 +eval/num steps total 157778 +eval/num paths total 158 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.010334 +eval/Actions Std 0.902496 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7293 +time/logging (s) 0.00435385 +time/sampling batch (s) 0.278736 +time/saving (s) 0.00410496 +time/training (s) 6.68701 +time/epoch (s) 9.7035 +time/total (s) 1682.81 +Epoch -843 +---------------------------------- --------------- +2022-05-10 13:38:49.182729 PDT | [1] Epoch -842 finished +---------------------------------- --------------- +epoch -842 +replay_buffer/size 999033 +trainer/num train calls 159000 +trainer/Policy Loss -19.6099 +trainer/Log Pis Mean 24.2461 +trainer/Log Pis Std 13.4435 +trainer/Log Pis Max 67.7897 +trainer/Log Pis Min -5.1372 +trainer/policy/mean Mean -0.0408373 +trainer/policy/mean Std 0.907446 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79744 +trainer/policy/normal/std Std 0.547763 +trainer/policy/normal/std Max 5.02361 +trainer/policy/normal/std Min 0.392602 +trainer/policy/normal/log_std Mean 1.00232 +trainer/policy/normal/log_std Std 0.25515 +trainer/policy/normal/log_std Max 1.61415 +trainer/policy/normal/log_std Min -0.93496 +eval/num steps total 158778 +eval/num paths total 159 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12281 +eval/Actions Std 0.908464 +eval/Actions Max 0.999978 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8665 +time/logging (s) 0.00379866 +time/sampling batch (s) 0.532246 +time/saving (s) 0.00343002 +time/training (s) 6.83411 +time/epoch (s) 10.2401 +time/total (s) 1693.05 +Epoch -842 +---------------------------------- --------------- +2022-05-10 13:38:59.321367 PDT | [1] Epoch -841 finished +---------------------------------- --------------- +epoch -841 +replay_buffer/size 999033 +trainer/num train calls 160000 +trainer/Policy Loss -19.1693 +trainer/Log Pis Mean 24.2945 +trainer/Log Pis Std 12.7736 +trainer/Log Pis Max 64.6637 +trainer/Log Pis Min -8.44549 +trainer/policy/mean Mean -0.0437699 +trainer/policy/mean Std 0.905416 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85545 +trainer/policy/normal/std Std 0.601244 +trainer/policy/normal/std Max 5.91888 +trainer/policy/normal/std Min 0.398129 +trainer/policy/normal/log_std Mean 1.01891 +trainer/policy/normal/log_std Std 0.273147 +trainer/policy/normal/log_std Max 1.77815 +trainer/policy/normal/log_std Min -0.92098 +eval/num steps total 159778 +eval/num paths total 160 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.072359 +eval/Actions Std 0.921569 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58715 +time/logging (s) 0.00380772 +time/sampling batch (s) 0.276672 +time/saving (s) 0.00342432 +time/training (s) 7.24727 +time/epoch (s) 10.1183 +time/total (s) 1703.17 +Epoch -841 +---------------------------------- --------------- +2022-05-10 13:39:09.783293 PDT | [1] Epoch -840 finished +---------------------------------- --------------- +epoch -840 +replay_buffer/size 999033 +trainer/num train calls 161000 +trainer/Policy Loss -19.2021 +trainer/Log Pis Mean 23.6346 +trainer/Log Pis Std 12.7687 +trainer/Log Pis Max 60.7979 +trainer/Log Pis Min -8.00156 +trainer/policy/mean Mean -0.0365961 +trainer/policy/mean Std 0.904343 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85457 +trainer/policy/normal/std Std 0.569503 +trainer/policy/normal/std Max 4.98666 +trainer/policy/normal/std Min 0.494568 +trainer/policy/normal/log_std Mean 1.02181 +trainer/policy/normal/log_std Std 0.257091 +trainer/policy/normal/log_std Max 1.60677 +trainer/policy/normal/log_std Min -0.70407 +eval/num steps total 160778 +eval/num paths total 161 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.241972 +eval/Actions Std 0.840871 +eval/Actions Max 0.999982 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68846 +time/logging (s) 0.00408327 +time/sampling batch (s) 0.533105 +time/saving (s) 0.00391286 +time/training (s) 7.21202 +time/epoch (s) 10.4416 +time/total (s) 1713.62 +Epoch -840 +---------------------------------- --------------- +2022-05-10 13:39:19.719653 PDT | [1] Epoch -839 finished +---------------------------------- --------------- +epoch -839 +replay_buffer/size 999033 +trainer/num train calls 162000 +trainer/Policy Loss -18.8389 +trainer/Log Pis Mean 24.1092 +trainer/Log Pis Std 12.7768 +trainer/Log Pis Max 69.503 +trainer/Log Pis Min -6.79791 +trainer/policy/mean Mean -0.0273191 +trainer/policy/mean Std 0.906507 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84956 +trainer/policy/normal/std Std 0.568471 +trainer/policy/normal/std Max 5.54894 +trainer/policy/normal/std Min 0.429395 +trainer/policy/normal/log_std Mean 1.01988 +trainer/policy/normal/log_std Std 0.259176 +trainer/policy/normal/log_std Max 1.71361 +trainer/policy/normal/log_std Min -0.845379 +eval/num steps total 161778 +eval/num paths total 162 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.284904 +eval/Actions Std 0.874642 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86781 +time/logging (s) 0.00369378 +time/sampling batch (s) 0.275646 +time/saving (s) 0.00339554 +time/training (s) 6.76479 +time/epoch (s) 9.91533 +time/total (s) 1723.54 +Epoch -839 +---------------------------------- --------------- +2022-05-10 13:39:30.305748 PDT | [1] Epoch -838 finished +---------------------------------- --------------- +epoch -838 +replay_buffer/size 999033 +trainer/num train calls 163000 +trainer/Policy Loss -19.9158 +trainer/Log Pis Mean 24.5453 +trainer/Log Pis Std 13.4251 +trainer/Log Pis Max 69.6167 +trainer/Log Pis Min -8.18186 +trainer/policy/mean Mean -0.0561619 +trainer/policy/mean Std 0.899567 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86306 +trainer/policy/normal/std Std 0.597818 +trainer/policy/normal/std Max 5.39179 +trainer/policy/normal/std Min 0.3803 +trainer/policy/normal/log_std Mean 1.02148 +trainer/policy/normal/log_std Std 0.274565 +trainer/policy/normal/log_std Max 1.68488 +trainer/policy/normal/log_std Min -0.966794 +eval/num steps total 162778 +eval/num paths total 163 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00379762 +eval/Actions Std 0.900646 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38083 +time/logging (s) 0.00384255 +time/sampling batch (s) 0.277387 +time/saving (s) 0.00345066 +time/training (s) 7.90036 +time/epoch (s) 10.5659 +time/total (s) 1734.11 +Epoch -838 +---------------------------------- --------------- +2022-05-10 13:39:40.236133 PDT | [1] Epoch -837 finished +---------------------------------- --------------- +epoch -837 +replay_buffer/size 999033 +trainer/num train calls 164000 +trainer/Policy Loss -20.2461 +trainer/Log Pis Mean 23.1025 +trainer/Log Pis Std 12.912 +trainer/Log Pis Max 61.9238 +trainer/Log Pis Min -7.9716 +trainer/policy/mean Mean -0.0203735 +trainer/policy/mean Std 0.905662 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.87445 +trainer/policy/normal/std Std 0.603281 +trainer/policy/normal/std Max 5.37978 +trainer/policy/normal/std Min 0.390425 +trainer/policy/normal/log_std Mean 1.02478 +trainer/policy/normal/log_std Std 0.278956 +trainer/policy/normal/log_std Max 1.68265 +trainer/policy/normal/log_std Min -0.940518 +eval/num steps total 163778 +eval/num paths total 164 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0409907 +eval/Actions Std 0.896358 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86387 +time/logging (s) 0.00379825 +time/sampling batch (s) 0.277962 +time/saving (s) 0.00337769 +time/training (s) 6.76096 +time/epoch (s) 9.90997 +time/total (s) 1744.02 +Epoch -837 +---------------------------------- --------------- +2022-05-10 13:39:52.259495 PDT | [1] Epoch -836 finished +---------------------------------- --------------- +epoch -836 +replay_buffer/size 999033 +trainer/num train calls 165000 +trainer/Policy Loss -17.5371 +trainer/Log Pis Mean 23.8776 +trainer/Log Pis Std 12.8224 +trainer/Log Pis Max 70.2179 +trainer/Log Pis Min -9.18773 +trainer/policy/mean Mean -0.0379033 +trainer/policy/mean Std 0.901766 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.79633 +trainer/policy/normal/std Std 0.581565 +trainer/policy/normal/std Max 4.93087 +trainer/policy/normal/std Min 0.456989 +trainer/policy/normal/log_std Mean 0.997805 +trainer/policy/normal/log_std Std 0.275292 +trainer/policy/normal/log_std Max 1.59552 +trainer/policy/normal/log_std Min -0.783095 +eval/num steps total 164778 +eval/num paths total 165 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.230869 +eval/Actions Std 0.876204 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5803 +time/logging (s) 0.00422923 +time/sampling batch (s) 0.530034 +time/saving (s) 0.0037965 +time/training (s) 8.88481 +time/epoch (s) 12.0032 +time/total (s) 1756.03 +Epoch -836 +---------------------------------- --------------- +2022-05-10 13:40:01.508456 PDT | [1] Epoch -835 finished +---------------------------------- --------------- +epoch -835 +replay_buffer/size 999033 +trainer/num train calls 166000 +trainer/Policy Loss -20.1652 +trainer/Log Pis Mean 23.6039 +trainer/Log Pis Std 13.114 +trainer/Log Pis Max 74.3584 +trainer/Log Pis Min -8.08828 +trainer/policy/mean Mean -0.0346322 +trainer/policy/mean Std 0.906466 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83983 +trainer/policy/normal/std Std 0.557774 +trainer/policy/normal/std Max 5.30263 +trainer/policy/normal/std Min 0.449487 +trainer/policy/normal/log_std Mean 1.01829 +trainer/policy/normal/log_std Std 0.247171 +trainer/policy/normal/log_std Max 1.6682 +trainer/policy/normal/log_std Min -0.799649 +eval/num steps total 165778 +eval/num paths total 166 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109147 +eval/Actions Std 0.907097 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57943 +time/logging (s) 0.00391122 +time/sampling batch (s) 0.278961 +time/saving (s) 0.00379822 +time/training (s) 6.36146 +time/epoch (s) 9.22755 +time/total (s) 1765.26 +Epoch -835 +---------------------------------- --------------- +2022-05-10 13:40:11.537119 PDT | [1] Epoch -834 finished +---------------------------------- --------------- +epoch -834 +replay_buffer/size 999033 +trainer/num train calls 167000 +trainer/Policy Loss -20.08 +trainer/Log Pis Mean 23.7465 +trainer/Log Pis Std 12.7236 +trainer/Log Pis Max 55.0751 +trainer/Log Pis Min -10.8874 +trainer/policy/mean Mean -0.0389434 +trainer/policy/mean Std 0.904761 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.90634 +trainer/policy/normal/std Std 0.610471 +trainer/policy/normal/std Max 6.53759 +trainer/policy/normal/std Min 0.366614 +trainer/policy/normal/log_std Mean 1.03606 +trainer/policy/normal/log_std Std 0.277165 +trainer/policy/normal/log_std Max 1.87757 +trainer/policy/normal/log_std Min -1.00345 +eval/num steps total 166778 +eval/num paths total 167 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0378425 +eval/Actions Std 0.905607 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79365 +time/logging (s) 0.00378999 +time/sampling batch (s) 0.285165 +time/saving (s) 0.00346819 +time/training (s) 6.92142 +time/epoch (s) 10.0075 +time/total (s) 1775.27 +Epoch -834 +---------------------------------- --------------- +2022-05-10 13:40:21.310727 PDT | [1] Epoch -833 finished +---------------------------------- --------------- +epoch -833 +replay_buffer/size 999033 +trainer/num train calls 168000 +trainer/Policy Loss -18.8457 +trainer/Log Pis Mean 24.5783 +trainer/Log Pis Std 12.9973 +trainer/Log Pis Max 68.7552 +trainer/Log Pis Min -6.9683 +trainer/policy/mean Mean -0.0437795 +trainer/policy/mean Std 0.902019 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88615 +trainer/policy/normal/std Std 0.618054 +trainer/policy/normal/std Max 5.43368 +trainer/policy/normal/std Min 0.427436 +trainer/policy/normal/log_std Mean 1.02804 +trainer/policy/normal/log_std Std 0.2813 +trainer/policy/normal/log_std Max 1.69262 +trainer/policy/normal/log_std Min -0.849951 +eval/num steps total 167778 +eval/num paths total 168 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0531529 +eval/Actions Std 0.893994 +eval/Actions Max 0.999989 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46374 +time/logging (s) 0.00374898 +time/sampling batch (s) 0.276057 +time/saving (s) 0.00340714 +time/training (s) 7.00628 +time/epoch (s) 9.75324 +time/total (s) 1785.02 +Epoch -833 +---------------------------------- --------------- +2022-05-10 13:40:32.412896 PDT | [1] Epoch -832 finished +---------------------------------- --------------- +epoch -832 +replay_buffer/size 999033 +trainer/num train calls 169000 +trainer/Policy Loss -18.9818 +trainer/Log Pis Mean 24.072 +trainer/Log Pis Std 13.3121 +trainer/Log Pis Max 66.986 +trainer/Log Pis Min -8.78079 +trainer/policy/mean Mean -0.0276755 +trainer/policy/mean Std 0.9062 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84823 +trainer/policy/normal/std Std 0.574306 +trainer/policy/normal/std Max 5.6943 +trainer/policy/normal/std Min 0.379951 +trainer/policy/normal/log_std Mean 1.01791 +trainer/policy/normal/log_std Std 0.268256 +trainer/policy/normal/log_std Max 1.73947 +trainer/policy/normal/log_std Min -0.967713 +eval/num steps total 168778 +eval/num paths total 169 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0837193 +eval/Actions Std 0.901168 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54742 +time/logging (s) 0.00373597 +time/sampling batch (s) 0.526909 +time/saving (s) 0.00343793 +time/training (s) 8.00027 +time/epoch (s) 11.0818 +time/total (s) 1796.11 +Epoch -832 +---------------------------------- --------------- +2022-05-10 13:40:43.680580 PDT | [1] Epoch -831 finished +---------------------------------- --------------- +epoch -831 +replay_buffer/size 999033 +trainer/num train calls 170000 +trainer/Policy Loss -19.1841 +trainer/Log Pis Mean 24.6284 +trainer/Log Pis Std 13.1991 +trainer/Log Pis Max 65.5115 +trainer/Log Pis Min -12.8266 +trainer/policy/mean Mean -0.0235565 +trainer/policy/mean Std 0.905038 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85549 +trainer/policy/normal/std Std 0.576029 +trainer/policy/normal/std Max 6.48475 +trainer/policy/normal/std Min 0.43244 +trainer/policy/normal/log_std Mean 1.02077 +trainer/policy/normal/log_std Std 0.266273 +trainer/policy/normal/log_std Max 1.86945 +trainer/policy/normal/log_std Min -0.838311 +eval/num steps total 169778 +eval/num paths total 170 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0623743 +eval/Actions Std 0.908222 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76694 +time/logging (s) 0.00385295 +time/sampling batch (s) 0.776332 +time/saving (s) 0.00343343 +time/training (s) 7.69706 +time/epoch (s) 11.2476 +time/total (s) 1807.36 +Epoch -831 +---------------------------------- --------------- +2022-05-10 13:40:55.733161 PDT | [1] Epoch -830 finished +---------------------------------- --------------- +epoch -830 +replay_buffer/size 999033 +trainer/num train calls 171000 +trainer/Policy Loss -19.9273 +trainer/Log Pis Mean 25.8347 +trainer/Log Pis Std 14.2531 +trainer/Log Pis Max 68.1923 +trainer/Log Pis Min -9.92076 +trainer/policy/mean Mean -0.0409077 +trainer/policy/mean Std 0.908933 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86697 +trainer/policy/normal/std Std 0.602366 +trainer/policy/normal/std Max 5.31493 +trainer/policy/normal/std Min 0.422364 +trainer/policy/normal/log_std Mean 1.02165 +trainer/policy/normal/log_std Std 0.282407 +trainer/policy/normal/log_std Max 1.67052 +trainer/policy/normal/log_std Min -0.861887 +eval/num steps total 170778 +eval/num paths total 171 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10141 +eval/Actions Std 0.916444 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76743 +time/logging (s) 0.0043246 +time/sampling batch (s) 0.777783 +time/saving (s) 0.00413529 +time/training (s) 8.47903 +time/epoch (s) 12.0327 +time/total (s) 1819.39 +Epoch -830 +---------------------------------- --------------- +2022-05-10 13:41:06.634491 PDT | [1] Epoch -829 finished +---------------------------------- --------------- +epoch -829 +replay_buffer/size 999033 +trainer/num train calls 172000 +trainer/Policy Loss -19.8788 +trainer/Log Pis Mean 25.2214 +trainer/Log Pis Std 13.7502 +trainer/Log Pis Max 76.1445 +trainer/Log Pis Min -13.1607 +trainer/policy/mean Mean -0.0408151 +trainer/policy/mean Std 0.909766 +trainer/policy/mean Max 0.999972 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86947 +trainer/policy/normal/std Std 0.574529 +trainer/policy/normal/std Max 5.5548 +trainer/policy/normal/std Min 0.530076 +trainer/policy/normal/log_std Mean 1.02748 +trainer/policy/normal/log_std Std 0.253232 +trainer/policy/normal/log_std Max 1.71466 +trainer/policy/normal/log_std Min -0.634734 +eval/num steps total 171778 +eval/num paths total 172 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.26643 +eval/Actions Std 0.865469 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76305 +time/logging (s) 0.00378733 +time/sampling batch (s) 0.277227 +time/saving (s) 0.00355987 +time/training (s) 7.83214 +time/epoch (s) 10.8798 +time/total (s) 1830.28 +Epoch -829 +---------------------------------- --------------- +2022-05-10 13:41:15.967213 PDT | [1] Epoch -828 finished +---------------------------------- --------------- +epoch -828 +replay_buffer/size 999033 +trainer/num train calls 173000 +trainer/Policy Loss -18.1587 +trainer/Log Pis Mean 24.9116 +trainer/Log Pis Std 13.2392 +trainer/Log Pis Max 62.069 +trainer/Log Pis Min -9.88365 +trainer/policy/mean Mean -0.0499007 +trainer/policy/mean Std 0.902501 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89337 +trainer/policy/normal/std Std 0.589343 +trainer/policy/normal/std Max 5.4128 +trainer/policy/normal/std Min 0.448471 +trainer/policy/normal/log_std Mean 1.03387 +trainer/policy/normal/log_std Std 0.265396 +trainer/policy/normal/log_std Max 1.68877 +trainer/policy/normal/log_std Min -0.801912 +eval/num steps total 172778 +eval/num paths total 173 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0963757 +eval/Actions Std 0.856304 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30208 +time/logging (s) 0.00383531 +time/sampling batch (s) 0.276588 +time/saving (s) 0.00367205 +time/training (s) 6.72642 +time/epoch (s) 9.31259 +time/total (s) 1839.59 +Epoch -828 +---------------------------------- --------------- +2022-05-10 13:41:26.582680 PDT | [1] Epoch -827 finished +---------------------------------- --------------- +epoch -827 +replay_buffer/size 999033 +trainer/num train calls 174000 +trainer/Policy Loss -18.9199 +trainer/Log Pis Mean 24.188 +trainer/Log Pis Std 13.7978 +trainer/Log Pis Max 69.5664 +trainer/Log Pis Min -12.0281 +trainer/policy/mean Mean -0.0342972 +trainer/policy/mean Std 0.901465 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86827 +trainer/policy/normal/std Std 0.582859 +trainer/policy/normal/std Max 6.70577 +trainer/policy/normal/std Min 0.443171 +trainer/policy/normal/log_std Mean 1.026 +trainer/policy/normal/log_std Std 0.259712 +trainer/policy/normal/log_std Max 1.90297 +trainer/policy/normal/log_std Min -0.8138 +eval/num steps total 173778 +eval/num paths total 174 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0162805 +eval/Actions Std 0.90377 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57498 +time/logging (s) 0.0037673 +time/sampling batch (s) 0.275848 +time/saving (s) 0.00352145 +time/training (s) 7.73688 +time/epoch (s) 10.595 +time/total (s) 1850.19 +Epoch -827 +---------------------------------- --------------- +2022-05-10 13:41:37.151499 PDT | [1] Epoch -826 finished +---------------------------------- --------------- +epoch -826 +replay_buffer/size 999033 +trainer/num train calls 175000 +trainer/Policy Loss -18.745 +trainer/Log Pis Mean 23.2034 +trainer/Log Pis Std 13.3318 +trainer/Log Pis Max 69.1599 +trainer/Log Pis Min -9.73027 +trainer/policy/mean Mean -0.0127933 +trainer/policy/mean Std 0.905404 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84181 +trainer/policy/normal/std Std 0.573313 +trainer/policy/normal/std Max 5.74657 +trainer/policy/normal/std Min 0.453978 +trainer/policy/normal/log_std Mean 1.01664 +trainer/policy/normal/log_std Std 0.260582 +trainer/policy/normal/log_std Max 1.7486 +trainer/policy/normal/log_std Min -0.789707 +eval/num steps total 174778 +eval/num paths total 175 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102068 +eval/Actions Std 0.913814 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46486 +time/logging (s) 0.00403275 +time/sampling batch (s) 0.776844 +time/saving (s) 0.00352924 +time/training (s) 7.29949 +time/epoch (s) 10.5488 +time/total (s) 1860.74 +Epoch -826 +---------------------------------- --------------- +2022-05-10 13:41:47.421232 PDT | [1] Epoch -825 finished +---------------------------------- --------------- +epoch -825 +replay_buffer/size 999033 +trainer/num train calls 176000 +trainer/Policy Loss -18.4816 +trainer/Log Pis Mean 24.3523 +trainer/Log Pis Std 12.8707 +trainer/Log Pis Max 68.3015 +trainer/Log Pis Min -7.0111 +trainer/policy/mean Mean -0.0309535 +trainer/policy/mean Std 0.906203 +trainer/policy/mean Max 0.999965 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.7809 +trainer/policy/normal/std Std 0.574222 +trainer/policy/normal/std Max 4.97194 +trainer/policy/normal/std Min 0.427006 +trainer/policy/normal/log_std Mean 0.992744 +trainer/policy/normal/log_std Std 0.273659 +trainer/policy/normal/log_std Max 1.60381 +trainer/policy/normal/log_std Min -0.850958 +eval/num steps total 175778 +eval/num paths total 176 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103189 +eval/Actions Std 0.916172 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47566 +time/logging (s) 0.00416277 +time/sampling batch (s) 0.527116 +time/saving (s) 0.00383793 +time/training (s) 7.23862 +time/epoch (s) 10.2494 +time/total (s) 1870.99 +Epoch -825 +---------------------------------- --------------- +2022-05-10 13:41:57.710483 PDT | [1] Epoch -824 finished +---------------------------------- --------------- +epoch -824 +replay_buffer/size 999033 +trainer/num train calls 177000 +trainer/Policy Loss -20.3772 +trainer/Log Pis Mean 24.3672 +trainer/Log Pis Std 13.2586 +trainer/Log Pis Max 81.2517 +trainer/Log Pis Min -3.84054 +trainer/policy/mean Mean -0.032236 +trainer/policy/mean Std 0.908064 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90057 +trainer/policy/normal/std Std 0.588966 +trainer/policy/normal/std Max 5.39549 +trainer/policy/normal/std Min 0.432975 +trainer/policy/normal/log_std Mean 1.03689 +trainer/policy/normal/log_std Std 0.261519 +trainer/policy/normal/log_std Max 1.68556 +trainer/policy/normal/log_std Min -0.837076 +eval/num steps total 176778 +eval/num paths total 177 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.016084 +eval/Actions Std 0.910613 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4863 +time/logging (s) 0.00388492 +time/sampling batch (s) 0.276802 +time/saving (s) 0.00368848 +time/training (s) 7.49743 +time/epoch (s) 10.2681 +time/total (s) 1881.27 +Epoch -824 +---------------------------------- --------------- +2022-05-10 13:42:09.026591 PDT | [1] Epoch -823 finished +---------------------------------- --------------- +epoch -823 +replay_buffer/size 999033 +trainer/num train calls 178000 +trainer/Policy Loss -18.9679 +trainer/Log Pis Mean 23.9231 +trainer/Log Pis Std 12.82 +trainer/Log Pis Max 73.3446 +trainer/Log Pis Min -5.42122 +trainer/policy/mean Mean -0.019324 +trainer/policy/mean Std 0.905331 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84518 +trainer/policy/normal/std Std 0.575713 +trainer/policy/normal/std Max 5.02058 +trainer/policy/normal/std Min 0.375974 +trainer/policy/normal/log_std Mean 1.0171 +trainer/policy/normal/log_std Std 0.266235 +trainer/policy/normal/log_std Max 1.61355 +trainer/policy/normal/log_std Min -0.978236 +eval/num steps total 177778 +eval/num paths total 178 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0895519 +eval/Actions Std 0.913682 +eval/Actions Max 0.999989 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62861 +time/logging (s) 0.00374512 +time/sampling batch (s) 0.526421 +time/saving (s) 0.00343558 +time/training (s) 8.13327 +time/epoch (s) 11.2955 +time/total (s) 1892.56 +Epoch -823 +---------------------------------- --------------- +2022-05-10 13:42:19.022124 PDT | [1] Epoch -822 finished +---------------------------------- --------------- +epoch -822 +replay_buffer/size 999033 +trainer/num train calls 179000 +trainer/Policy Loss -19.7805 +trainer/Log Pis Mean 24.7094 +trainer/Log Pis Std 12.9407 +trainer/Log Pis Max 72.8869 +trainer/Log Pis Min -5.68799 +trainer/policy/mean Mean -0.0332562 +trainer/policy/mean Std 0.904922 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86629 +trainer/policy/normal/std Std 0.576067 +trainer/policy/normal/std Max 5.40849 +trainer/policy/normal/std Min 0.386639 +trainer/policy/normal/log_std Mean 1.02496 +trainer/policy/normal/log_std Std 0.263597 +trainer/policy/normal/log_std Max 1.68797 +trainer/policy/normal/log_std Min -0.950265 +eval/num steps total 178778 +eval/num paths total 179 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.39464 +eval/Actions Std 0.885262 +eval/Actions Max 0.999991 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56367 +time/logging (s) 0.00375905 +time/sampling batch (s) 0.275491 +time/saving (s) 0.00344025 +time/training (s) 7.12897 +time/epoch (s) 9.97533 +time/total (s) 1902.54 +Epoch -822 +---------------------------------- --------------- +2022-05-10 13:42:29.724693 PDT | [1] Epoch -821 finished +---------------------------------- --------------- +epoch -821 +replay_buffer/size 999033 +trainer/num train calls 180000 +trainer/Policy Loss -19.4552 +trainer/Log Pis Mean 24.8158 +trainer/Log Pis Std 13.488 +trainer/Log Pis Max 71.3675 +trainer/Log Pis Min -9.22894 +trainer/policy/mean Mean -0.0562699 +trainer/policy/mean Std 0.907393 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8616 +trainer/policy/normal/std Std 0.609614 +trainer/policy/normal/std Max 5.2787 +trainer/policy/normal/std Min 0.442576 +trainer/policy/normal/log_std Mean 1.01929 +trainer/policy/normal/log_std Std 0.283212 +trainer/policy/normal/log_std Max 1.66368 +trainer/policy/normal/log_std Min -0.815143 +eval/num steps total 179778 +eval/num paths total 180 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0609893 +eval/Actions Std 0.909418 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59359 +time/logging (s) 0.0037708 +time/sampling batch (s) 0.318724 +time/saving (s) 0.00365746 +time/training (s) 7.76206 +time/epoch (s) 10.6818 +time/total (s) 1913.23 +Epoch -821 +---------------------------------- --------------- +2022-05-10 13:42:40.180485 PDT | [1] Epoch -820 finished +---------------------------------- --------------- +epoch -820 +replay_buffer/size 999033 +trainer/num train calls 181000 +trainer/Policy Loss -19.6872 +trainer/Log Pis Mean 24.378 +trainer/Log Pis Std 13.0667 +trainer/Log Pis Max 62.2981 +trainer/Log Pis Min -8.68156 +trainer/policy/mean Mean -0.0484404 +trainer/policy/mean Std 0.905473 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81057 +trainer/policy/normal/std Std 0.583212 +trainer/policy/normal/std Max 5.08784 +trainer/policy/normal/std Min 0.369908 +trainer/policy/normal/log_std Mean 1.0015 +trainer/policy/normal/log_std Std 0.286234 +trainer/policy/normal/log_std Max 1.62685 +trainer/policy/normal/log_std Min -0.994501 +eval/num steps total 180778 +eval/num paths total 181 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102301 +eval/Actions Std 0.917281 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77969 +time/logging (s) 0.0040527 +time/sampling batch (s) 0.553793 +time/saving (s) 0.00381576 +time/training (s) 7.09393 +time/epoch (s) 10.4353 +time/total (s) 1923.67 +Epoch -820 +---------------------------------- --------------- +2022-05-10 13:42:49.621309 PDT | [1] Epoch -819 finished +---------------------------------- --------------- +epoch -819 +replay_buffer/size 999033 +trainer/num train calls 182000 +trainer/Policy Loss -18.0152 +trainer/Log Pis Mean 23.9782 +trainer/Log Pis Std 13.0418 +trainer/Log Pis Max 64.8835 +trainer/Log Pis Min -8.34168 +trainer/policy/mean Mean -0.0301246 +trainer/policy/mean Std 0.903968 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85214 +trainer/policy/normal/std Std 0.594099 +trainer/policy/normal/std Max 5.25943 +trainer/policy/normal/std Min 0.432004 +trainer/policy/normal/log_std Mean 1.01831 +trainer/policy/normal/log_std Std 0.270435 +trainer/policy/normal/log_std Max 1.66002 +trainer/policy/normal/log_std Min -0.83932 +eval/num steps total 181778 +eval/num paths total 182 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0213237 +eval/Actions Std 0.901627 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44773 +time/logging (s) 0.00379144 +time/sampling batch (s) 0.275113 +time/saving (s) 0.00369029 +time/training (s) 6.6895 +time/epoch (s) 9.41983 +time/total (s) 1933.09 +Epoch -819 +---------------------------------- --------------- +2022-05-10 13:43:00.940413 PDT | [1] Epoch -818 finished +---------------------------------- --------------- +epoch -818 +replay_buffer/size 999033 +trainer/num train calls 183000 +trainer/Policy Loss -19.2616 +trainer/Log Pis Mean 24.236 +trainer/Log Pis Std 13.4537 +trainer/Log Pis Max 69.5772 +trainer/Log Pis Min -10.4684 +trainer/policy/mean Mean -0.034105 +trainer/policy/mean Std 0.902604 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82752 +trainer/policy/normal/std Std 0.580904 +trainer/policy/normal/std Max 5.38297 +trainer/policy/normal/std Min 0.376712 +trainer/policy/normal/log_std Mean 1.0103 +trainer/policy/normal/log_std Std 0.267764 +trainer/policy/normal/log_std Max 1.68324 +trainer/policy/normal/log_std Min -0.976275 +eval/num steps total 182778 +eval/num paths total 183 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113629 +eval/Actions Std 0.919514 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54463 +time/logging (s) 0.00383812 +time/sampling batch (s) 0.781269 +time/saving (s) 0.00348497 +time/training (s) 7.96516 +time/epoch (s) 11.2984 +time/total (s) 1944.39 +Epoch -818 +---------------------------------- --------------- +2022-05-10 13:43:12.222701 PDT | [1] Epoch -817 finished +---------------------------------- --------------- +epoch -817 +replay_buffer/size 999033 +trainer/num train calls 184000 +trainer/Policy Loss -20.1037 +trainer/Log Pis Mean 24.7902 +trainer/Log Pis Std 13.6074 +trainer/Log Pis Max 80.9962 +trainer/Log Pis Min -4.41684 +trainer/policy/mean Mean -0.0443473 +trainer/policy/mean Std 0.90265 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83203 +trainer/policy/normal/std Std 0.579251 +trainer/policy/normal/std Max 5.70255 +trainer/policy/normal/std Min 0.435897 +trainer/policy/normal/log_std Mean 1.01188 +trainer/policy/normal/log_std Std 0.268493 +trainer/policy/normal/log_std Max 1.74091 +trainer/policy/normal/log_std Min -0.830348 +eval/num steps total 183778 +eval/num paths total 184 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124701 +eval/Actions Std 0.887341 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70678 +time/logging (s) 0.00438072 +time/sampling batch (s) 0.529342 +time/saving (s) 0.00355035 +time/training (s) 8.01824 +time/epoch (s) 11.2623 +time/total (s) 1955.66 +Epoch -817 +---------------------------------- --------------- +2022-05-10 13:43:22.926464 PDT | [1] Epoch -816 finished +---------------------------------- --------------- +epoch -816 +replay_buffer/size 999033 +trainer/num train calls 185000 +trainer/Policy Loss -19.8476 +trainer/Log Pis Mean 24.4365 +trainer/Log Pis Std 13.509 +trainer/Log Pis Max 66.0022 +trainer/Log Pis Min -10.0071 +trainer/policy/mean Mean -0.0503551 +trainer/policy/mean Std 0.910272 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8882 +trainer/policy/normal/std Std 0.570095 +trainer/policy/normal/std Max 5.93543 +trainer/policy/normal/std Min 0.447033 +trainer/policy/normal/log_std Mean 1.03448 +trainer/policy/normal/log_std Std 0.252695 +trainer/policy/normal/log_std Max 1.78094 +trainer/policy/normal/log_std Min -0.805124 +eval/num steps total 184778 +eval/num paths total 185 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0913902 +eval/Actions Std 0.899628 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59033 +time/logging (s) 0.00390174 +time/sampling batch (s) 0.526229 +time/saving (s) 0.00345791 +time/training (s) 7.55854 +time/epoch (s) 10.6825 +time/total (s) 1966.34 +Epoch -816 +---------------------------------- --------------- +2022-05-10 13:43:33.772917 PDT | [1] Epoch -815 finished +---------------------------------- --------------- +epoch -815 +replay_buffer/size 999033 +trainer/num train calls 186000 +trainer/Policy Loss -19.3213 +trainer/Log Pis Mean 24.0048 +trainer/Log Pis Std 13.3195 +trainer/Log Pis Max 66.6152 +trainer/Log Pis Min -7.16909 +trainer/policy/mean Mean -0.0241841 +trainer/policy/mean Std 0.898794 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85569 +trainer/policy/normal/std Std 0.591018 +trainer/policy/normal/std Max 5.48068 +trainer/policy/normal/std Min 0.364643 +trainer/policy/normal/log_std Mean 1.01827 +trainer/policy/normal/log_std Std 0.28044 +trainer/policy/normal/log_std Max 1.70123 +trainer/policy/normal/log_std Min -1.00884 +eval/num steps total 185574 +eval/num paths total 186 +eval/path length Mean 796 +eval/path length Std 0 +eval/path length Max 796 +eval/path length Min 796 +eval/Rewards Mean 0.00125628 +eval/Rewards Std 0.0354218 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0421371 +eval/Actions Std 0.896782 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.58158 +time/logging (s) 0.00330016 +time/sampling batch (s) 1.02748 +time/saving (s) 0.00339423 +time/training (s) 7.20966 +time/epoch (s) 10.8254 +time/total (s) 1977.17 +Epoch -815 +---------------------------------- --------------- +2022-05-10 13:43:43.954842 PDT | [1] Epoch -814 finished +---------------------------------- --------------- +epoch -814 +replay_buffer/size 999033 +trainer/num train calls 187000 +trainer/Policy Loss -19.35 +trainer/Log Pis Mean 24.8621 +trainer/Log Pis Std 13.1388 +trainer/Log Pis Max 64.7134 +trainer/Log Pis Min -11.2497 +trainer/policy/mean Mean -0.0374859 +trainer/policy/mean Std 0.906378 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82515 +trainer/policy/normal/std Std 0.627534 +trainer/policy/normal/std Max 5.45767 +trainer/policy/normal/std Min 0.410705 +trainer/policy/normal/log_std Mean 1.00386 +trainer/policy/normal/log_std Std 0.294618 +trainer/policy/normal/log_std Max 1.69702 +trainer/policy/normal/log_std Min -0.889881 +eval/num steps total 186574 +eval/num paths total 187 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0311685 +eval/Actions Std 0.908293 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62734 +time/logging (s) 0.00414754 +time/sampling batch (s) 0.275865 +time/saving (s) 0.0039968 +time/training (s) 7.25097 +time/epoch (s) 10.1623 +time/total (s) 1987.34 +Epoch -814 +---------------------------------- --------------- +2022-05-10 13:43:54.453750 PDT | [1] Epoch -813 finished +---------------------------------- --------------- +epoch -813 +replay_buffer/size 999033 +trainer/num train calls 188000 +trainer/Policy Loss -19.6542 +trainer/Log Pis Mean 23.6543 +trainer/Log Pis Std 13.5432 +trainer/Log Pis Max 72.264 +trainer/Log Pis Min -8.26917 +trainer/policy/mean Mean -0.0372914 +trainer/policy/mean Std 0.906114 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85802 +trainer/policy/normal/std Std 0.574594 +trainer/policy/normal/std Max 5.55282 +trainer/policy/normal/std Min 0.42328 +trainer/policy/normal/log_std Mean 1.02249 +trainer/policy/normal/log_std Std 0.261331 +trainer/policy/normal/log_std Max 1.71431 +trainer/policy/normal/log_std Min -0.859721 +eval/num steps total 187574 +eval/num paths total 188 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.367779 +eval/Actions Std 0.841426 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6162 +time/logging (s) 0.00373005 +time/sampling batch (s) 0.27593 +time/saving (s) 0.00344241 +time/training (s) 7.57843 +time/epoch (s) 10.4777 +time/total (s) 1997.82 +Epoch -813 +---------------------------------- --------------- +2022-05-10 13:44:05.016736 PDT | [1] Epoch -812 finished +---------------------------------- --------------- +epoch -812 +replay_buffer/size 999033 +trainer/num train calls 189000 +trainer/Policy Loss -19.8225 +trainer/Log Pis Mean 23.9932 +trainer/Log Pis Std 13.461 +trainer/Log Pis Max 81.4173 +trainer/Log Pis Min -6.96935 +trainer/policy/mean Mean -0.0198488 +trainer/policy/mean Std 0.905417 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.83009 +trainer/policy/normal/std Std 0.621208 +trainer/policy/normal/std Max 5.16122 +trainer/policy/normal/std Min 0.344323 +trainer/policy/normal/log_std Mean 1.00608 +trainer/policy/normal/log_std Std 0.293228 +trainer/policy/normal/log_std Max 1.64117 +trainer/policy/normal/log_std Min -1.06617 +eval/num steps total 188303 +eval/num paths total 189 +eval/path length Mean 729 +eval/path length Std 0 +eval/path length Max 729 +eval/path length Min 729 +eval/Rewards Mean 0.00137174 +eval/Rewards Std 0.0370116 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.00506502 +eval/Actions Std 0.903624 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.56342 +time/logging (s) 0.00297015 +time/sampling batch (s) 0.776673 +time/saving (s) 0.00341089 +time/training (s) 7.19543 +time/epoch (s) 10.5419 +time/total (s) 2008.36 +Epoch -812 +---------------------------------- --------------- +2022-05-10 13:44:15.489031 PDT | [1] Epoch -811 finished +---------------------------------- --------------- +epoch -811 +replay_buffer/size 999033 +trainer/num train calls 190000 +trainer/Policy Loss -19.4543 +trainer/Log Pis Mean 24.9633 +trainer/Log Pis Std 13.6145 +trainer/Log Pis Max 66.4763 +trainer/Log Pis Min -7.87376 +trainer/policy/mean Mean -0.0529537 +trainer/policy/mean Std 0.901135 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87498 +trainer/policy/normal/std Std 0.591491 +trainer/policy/normal/std Max 5.52458 +trainer/policy/normal/std Min 0.4201 +trainer/policy/normal/log_std Mean 1.02749 +trainer/policy/normal/log_std Std 0.263563 +trainer/policy/normal/log_std Max 1.70921 +trainer/policy/normal/log_std Min -0.867262 +eval/num steps total 189303 +eval/num paths total 190 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133315 +eval/Actions Std 0.896819 +eval/Actions Max 1 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57367 +time/logging (s) 0.00383245 +time/sampling batch (s) 0.526059 +time/saving (s) 0.00339899 +time/training (s) 7.34599 +time/epoch (s) 10.4529 +time/total (s) 2018.82 +Epoch -811 +---------------------------------- --------------- +2022-05-10 13:44:26.979378 PDT | [1] Epoch -810 finished +---------------------------------- --------------- +epoch -810 +replay_buffer/size 999033 +trainer/num train calls 191000 +trainer/Policy Loss -18.2327 +trainer/Log Pis Mean 23.379 +trainer/Log Pis Std 12.3149 +trainer/Log Pis Max 61.1995 +trainer/Log Pis Min -4.75153 +trainer/policy/mean Mean -0.0158189 +trainer/policy/mean Std 0.905832 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84883 +trainer/policy/normal/std Std 0.583638 +trainer/policy/normal/std Max 5.7102 +trainer/policy/normal/std Min 0.341725 +trainer/policy/normal/log_std Mean 1.01714 +trainer/policy/normal/log_std Std 0.273317 +trainer/policy/normal/log_std Max 1.74225 +trainer/policy/normal/log_std Min -1.07375 +eval/num steps total 190303 +eval/num paths total 191 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0513033 +eval/Actions Std 0.910637 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58557 +time/logging (s) 0.00380994 +time/sampling batch (s) 0.276428 +time/saving (s) 0.00348446 +time/training (s) 8.60063 +time/epoch (s) 11.4699 +time/total (s) 2030.29 +Epoch -810 +---------------------------------- --------------- +2022-05-10 13:44:37.818033 PDT | [1] Epoch -809 finished +---------------------------------- --------------- +epoch -809 +replay_buffer/size 999033 +trainer/num train calls 192000 +trainer/Policy Loss -19.8954 +trainer/Log Pis Mean 24.2525 +trainer/Log Pis Std 12.9985 +trainer/Log Pis Max 65.4718 +trainer/Log Pis Min -9.96444 +trainer/policy/mean Mean -0.0382418 +trainer/policy/mean Std 0.903539 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.86835 +trainer/policy/normal/std Std 0.587175 +trainer/policy/normal/std Max 5.21223 +trainer/policy/normal/std Min 0.397998 +trainer/policy/normal/log_std Mean 1.02518 +trainer/policy/normal/log_std Std 0.264937 +trainer/policy/normal/log_std Max 1.65101 +trainer/policy/normal/log_std Min -0.921309 +eval/num steps total 191303 +eval/num paths total 192 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.220339 +eval/Actions Std 0.900021 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66207 +time/logging (s) 0.00429584 +time/sampling batch (s) 1.03049 +time/saving (s) 0.00400168 +time/training (s) 7.11761 +time/epoch (s) 10.8185 +time/total (s) 2041.11 +Epoch -809 +---------------------------------- --------------- +2022-05-10 13:44:48.039201 PDT | [1] Epoch -808 finished +---------------------------------- --------------- +epoch -808 +replay_buffer/size 999033 +trainer/num train calls 193000 +trainer/Policy Loss -19.9265 +trainer/Log Pis Mean 24.5224 +trainer/Log Pis Std 12.5145 +trainer/Log Pis Max 65.3256 +trainer/Log Pis Min -9.14577 +trainer/policy/mean Mean -0.0402066 +trainer/policy/mean Std 0.905219 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78862 +trainer/policy/normal/std Std 0.592548 +trainer/policy/normal/std Max 5.71708 +trainer/policy/normal/std Min 0.369826 +trainer/policy/normal/log_std Mean 0.994105 +trainer/policy/normal/log_std Std 0.279391 +trainer/policy/normal/log_std Max 1.74346 +trainer/policy/normal/log_std Min -0.994723 +eval/num steps total 192303 +eval/num paths total 193 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123856 +eval/Actions Std 0.890152 +eval/Actions Max 0.99999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5736 +time/logging (s) 0.00387815 +time/sampling batch (s) 0.28121 +time/saving (s) 0.00360427 +time/training (s) 7.33713 +time/epoch (s) 10.1994 +time/total (s) 2051.31 +Epoch -808 +---------------------------------- --------------- +2022-05-10 13:44:57.784748 PDT | [1] Epoch -807 finished +---------------------------------- --------------- +epoch -807 +replay_buffer/size 999033 +trainer/num train calls 194000 +trainer/Policy Loss -19.1416 +trainer/Log Pis Mean 24.1377 +trainer/Log Pis Std 13.2677 +trainer/Log Pis Max 72.7104 +trainer/Log Pis Min -9.221 +trainer/policy/mean Mean -0.053148 +trainer/policy/mean Std 0.906916 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87189 +trainer/policy/normal/std Std 0.577165 +trainer/policy/normal/std Max 5.62029 +trainer/policy/normal/std Min 0.344379 +trainer/policy/normal/log_std Mean 1.02718 +trainer/policy/normal/log_std Std 0.262178 +trainer/policy/normal/log_std Max 1.72638 +trainer/policy/normal/log_std Min -1.06601 +eval/num steps total 193303 +eval/num paths total 194 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0363777 +eval/Actions Std 0.904117 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58941 +time/logging (s) 0.00387096 +time/sampling batch (s) 0.278434 +time/saving (s) 0.0034736 +time/training (s) 6.8497 +time/epoch (s) 9.72488 +time/total (s) 2061.04 +Epoch -807 +---------------------------------- --------------- +2022-05-10 13:45:07.833716 PDT | [1] Epoch -806 finished +---------------------------------- --------------- +epoch -806 +replay_buffer/size 999033 +trainer/num train calls 195000 +trainer/Policy Loss -18.5988 +trainer/Log Pis Mean 24.3524 +trainer/Log Pis Std 13.4824 +trainer/Log Pis Max 74.2293 +trainer/Log Pis Min -12.3419 +trainer/policy/mean Mean -0.0345586 +trainer/policy/mean Std 0.900911 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.8273 +trainer/policy/normal/std Std 0.613568 +trainer/policy/normal/std Max 5.65391 +trainer/policy/normal/std Min 0.283439 +trainer/policy/normal/log_std Mean 1.00549 +trainer/policy/normal/log_std Std 0.292749 +trainer/policy/normal/log_std Max 1.73235 +trainer/policy/normal/log_std Min -1.26076 +eval/num steps total 194303 +eval/num paths total 195 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.382111 +eval/Actions Std 0.864599 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71756 +time/logging (s) 0.00366703 +time/sampling batch (s) 0.526375 +time/saving (s) 0.00343641 +time/training (s) 6.77742 +time/epoch (s) 10.0285 +time/total (s) 2071.07 +Epoch -806 +---------------------------------- --------------- +2022-05-10 13:45:17.854001 PDT | [1] Epoch -805 finished +---------------------------------- --------------- +epoch -805 +replay_buffer/size 999033 +trainer/num train calls 196000 +trainer/Policy Loss -18.5994 +trainer/Log Pis Mean 24.6015 +trainer/Log Pis Std 13.3514 +trainer/Log Pis Max 64.3849 +trainer/Log Pis Min -7.60558 +trainer/policy/mean Mean -0.0113481 +trainer/policy/mean Std 0.90847 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85565 +trainer/policy/normal/std Std 0.566782 +trainer/policy/normal/std Max 4.90984 +trainer/policy/normal/std Min 0.463342 +trainer/policy/normal/log_std Mean 1.02213 +trainer/policy/normal/log_std Std 0.25888 +trainer/policy/normal/log_std Max 1.59124 +trainer/policy/normal/log_std Min -0.76929 +eval/num steps total 195303 +eval/num paths total 196 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.182176 +eval/Actions Std 0.920475 +eval/Actions Max 0.999994 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63968 +time/logging (s) 0.00368961 +time/sampling batch (s) 0.275468 +time/saving (s) 0.00342549 +time/training (s) 7.07762 +time/epoch (s) 9.99988 +time/total (s) 2081.08 +Epoch -805 +---------------------------------- --------------- +2022-05-10 13:45:28.582408 PDT | [1] Epoch -804 finished +---------------------------------- --------------- +epoch -804 +replay_buffer/size 999033 +trainer/num train calls 197000 +trainer/Policy Loss -20.1055 +trainer/Log Pis Mean 23.9525 +trainer/Log Pis Std 12.9652 +trainer/Log Pis Max 62.1877 +trainer/Log Pis Min -7.39243 +trainer/policy/mean Mean -0.061778 +trainer/policy/mean Std 0.903572 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86132 +trainer/policy/normal/std Std 0.605498 +trainer/policy/normal/std Max 5.21536 +trainer/policy/normal/std Min 0.44932 +trainer/policy/normal/log_std Mean 1.0198 +trainer/policy/normal/log_std Std 0.279925 +trainer/policy/normal/log_std Max 1.65161 +trainer/policy/normal/log_std Min -0.800019 +eval/num steps total 196303 +eval/num paths total 197 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0915124 +eval/Actions Std 0.919445 +eval/Actions Max 0.999977 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42003 +time/logging (s) 0.00371606 +time/sampling batch (s) 1.27707 +time/saving (s) 0.00341212 +time/training (s) 7.0037 +time/epoch (s) 10.7079 +time/total (s) 2091.79 +Epoch -804 +---------------------------------- --------------- +2022-05-10 13:45:39.735722 PDT | [1] Epoch -803 finished +---------------------------------- --------------- +epoch -803 +replay_buffer/size 999033 +trainer/num train calls 198000 +trainer/Policy Loss -19.246 +trainer/Log Pis Mean 24.1888 +trainer/Log Pis Std 13.4482 +trainer/Log Pis Max 66.5517 +trainer/Log Pis Min -7.45331 +trainer/policy/mean Mean -0.0483823 +trainer/policy/mean Std 0.906788 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84343 +trainer/policy/normal/std Std 0.584408 +trainer/policy/normal/std Max 6.14499 +trainer/policy/normal/std Min 0.358132 +trainer/policy/normal/log_std Mean 1.01543 +trainer/policy/normal/log_std Std 0.27201 +trainer/policy/normal/log_std Max 1.81564 +trainer/policy/normal/log_std Min -1.02685 +eval/num steps total 197303 +eval/num paths total 198 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0343861 +eval/Actions Std 0.90429 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70337 +time/logging (s) 0.00401791 +time/sampling batch (s) 0.777023 +time/saving (s) 0.00383074 +time/training (s) 7.64486 +time/epoch (s) 11.1331 +time/total (s) 2102.92 +Epoch -803 +---------------------------------- --------------- +2022-05-10 13:45:49.555243 PDT | [1] Epoch -802 finished +---------------------------------- --------------- +epoch -802 +replay_buffer/size 999033 +trainer/num train calls 199000 +trainer/Policy Loss -19.0397 +trainer/Log Pis Mean 24.2512 +trainer/Log Pis Std 12.2758 +trainer/Log Pis Max 64.2497 +trainer/Log Pis Min -9.81058 +trainer/policy/mean Mean -0.0262167 +trainer/policy/mean Std 0.902138 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83895 +trainer/policy/normal/std Std 0.580156 +trainer/policy/normal/std Max 5.51362 +trainer/policy/normal/std Min 0.393885 +trainer/policy/normal/log_std Mean 1.01495 +trainer/policy/normal/log_std Std 0.264535 +trainer/policy/normal/log_std Max 1.70722 +trainer/policy/normal/log_std Min -0.931697 +eval/num steps total 198303 +eval/num paths total 199 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00202 +eval/Actions Std 0.882332 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45278 +time/logging (s) 0.00380316 +time/sampling batch (s) 0.27369 +time/saving (s) 0.00357071 +time/training (s) 7.06498 +time/epoch (s) 9.79883 +time/total (s) 2112.72 +Epoch -802 +---------------------------------- --------------- +2022-05-10 13:45:59.563030 PDT | [1] Epoch -801 finished +---------------------------------- --------------- +epoch -801 +replay_buffer/size 999033 +trainer/num train calls 200000 +trainer/Policy Loss -19.4178 +trainer/Log Pis Mean 24.6469 +trainer/Log Pis Std 13.3267 +trainer/Log Pis Max 61.509 +trainer/Log Pis Min -7.14141 +trainer/policy/mean Mean -0.0141055 +trainer/policy/mean Std 0.90167 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87972 +trainer/policy/normal/std Std 0.619812 +trainer/policy/normal/std Max 5.60078 +trainer/policy/normal/std Min 0.440883 +trainer/policy/normal/log_std Mean 1.02525 +trainer/policy/normal/log_std Std 0.284546 +trainer/policy/normal/log_std Max 1.72291 +trainer/policy/normal/log_std Min -0.818976 +eval/num steps total 199303 +eval/num paths total 200 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0279269 +eval/Actions Std 0.90349 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56687 +time/logging (s) 0.00369613 +time/sampling batch (s) 0.523833 +time/saving (s) 0.00635134 +time/training (s) 6.88664 +time/epoch (s) 9.98739 +time/total (s) 2122.71 +Epoch -801 +---------------------------------- --------------- +2022-05-10 13:46:10.646535 PDT | [1] Epoch -800 finished +---------------------------------- --------------- +epoch -800 +replay_buffer/size 999033 +trainer/num train calls 201000 +trainer/Policy Loss -19.9148 +trainer/Log Pis Mean 24.4892 +trainer/Log Pis Std 12.8523 +trainer/Log Pis Max 60.7709 +trainer/Log Pis Min -14.6958 +trainer/policy/mean Mean -0.027271 +trainer/policy/mean Std 0.901594 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.85462 +trainer/policy/normal/std Std 0.60178 +trainer/policy/normal/std Max 5.83224 +trainer/policy/normal/std Min 0.420375 +trainer/policy/normal/log_std Mean 1.0188 +trainer/policy/normal/log_std Std 0.271529 +trainer/policy/normal/log_std Max 1.7634 +trainer/policy/normal/log_std Min -0.866607 +eval/num steps total 200303 +eval/num paths total 201 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0925916 +eval/Actions Std 0.862058 +eval/Actions Max 0.999994 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74562 +time/logging (s) 0.00365392 +time/sampling batch (s) 0.276896 +time/saving (s) 0.00338592 +time/training (s) 8.0334 +time/epoch (s) 11.063 +time/total (s) 2133.78 +Epoch -800 +---------------------------------- --------------- +2022-05-10 13:46:21.031850 PDT | [1] Epoch -799 finished +---------------------------------- --------------- +epoch -799 +replay_buffer/size 999033 +trainer/num train calls 202000 +trainer/Policy Loss -20.2215 +trainer/Log Pis Mean 25.12 +trainer/Log Pis Std 12.9194 +trainer/Log Pis Max 66.8245 +trainer/Log Pis Min -10.9589 +trainer/policy/mean Mean -0.0321504 +trainer/policy/mean Std 0.901533 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8276 +trainer/policy/normal/std Std 0.622084 +trainer/policy/normal/std Max 6.03971 +trainer/policy/normal/std Min 0.420506 +trainer/policy/normal/log_std Mean 1.00404 +trainer/policy/normal/log_std Std 0.30043 +trainer/policy/normal/log_std Max 1.79836 +trainer/policy/normal/log_std Min -0.866297 +eval/num steps total 201303 +eval/num paths total 202 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.213924 +eval/Actions Std 0.947561 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33711 +time/logging (s) 0.00373 +time/sampling batch (s) 0.280108 +time/saving (s) 0.00345845 +time/training (s) 7.74008 +time/epoch (s) 10.3645 +time/total (s) 2144.15 +Epoch -799 +---------------------------------- --------------- +2022-05-10 13:46:31.068797 PDT | [1] Epoch -798 finished +---------------------------------- --------------- +epoch -798 +replay_buffer/size 999033 +trainer/num train calls 203000 +trainer/Policy Loss -19.4319 +trainer/Log Pis Mean 24.5184 +trainer/Log Pis Std 12.3809 +trainer/Log Pis Max 69.2622 +trainer/Log Pis Min -1.74546 +trainer/policy/mean Mean -0.0189757 +trainer/policy/mean Std 0.904584 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81237 +trainer/policy/normal/std Std 0.602499 +trainer/policy/normal/std Max 5.38253 +trainer/policy/normal/std Min 0.377481 +trainer/policy/normal/log_std Mean 1.00171 +trainer/policy/normal/log_std Std 0.283565 +trainer/policy/normal/log_std Max 1.68316 +trainer/policy/normal/log_std Min -0.974234 +eval/num steps total 202303 +eval/num paths total 203 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00117414 +eval/Actions Std 0.903789 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50862 +time/logging (s) 0.00411234 +time/sampling batch (s) 0.279014 +time/saving (s) 0.00383109 +time/training (s) 7.221 +time/epoch (s) 10.0166 +time/total (s) 2154.17 +Epoch -798 +---------------------------------- --------------- +2022-05-10 13:46:40.553464 PDT | [1] Epoch -797 finished +---------------------------------- --------------- +epoch -797 +replay_buffer/size 999033 +trainer/num train calls 204000 +trainer/Policy Loss -19.453 +trainer/Log Pis Mean 25.2423 +trainer/Log Pis Std 12.9284 +trainer/Log Pis Max 62.4512 +trainer/Log Pis Min -8.42709 +trainer/policy/mean Mean -0.0578996 +trainer/policy/mean Std 0.90339 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.89237 +trainer/policy/normal/std Std 0.589007 +trainer/policy/normal/std Max 4.85322 +trainer/policy/normal/std Min 0.414701 +trainer/policy/normal/log_std Mean 1.03254 +trainer/policy/normal/log_std Std 0.272198 +trainer/policy/normal/log_std Max 1.57964 +trainer/policy/normal/log_std Min -0.880198 +eval/num steps total 203303 +eval/num paths total 204 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0944398 +eval/Actions Std 0.901975 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64485 +time/logging (s) 0.00394258 +time/sampling batch (s) 0.52888 +time/saving (s) 0.00376308 +time/training (s) 6.28188 +time/epoch (s) 9.46332 +time/total (s) 2163.63 +Epoch -797 +---------------------------------- --------------- +2022-05-10 13:46:50.272673 PDT | [1] Epoch -796 finished +---------------------------------- --------------- +epoch -796 +replay_buffer/size 999033 +trainer/num train calls 205000 +trainer/Policy Loss -19.0061 +trainer/Log Pis Mean 24.1034 +trainer/Log Pis Std 12.7877 +trainer/Log Pis Max 61.0173 +trainer/Log Pis Min -12.4996 +trainer/policy/mean Mean -0.0296698 +trainer/policy/mean Std 0.905578 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86365 +trainer/policy/normal/std Std 0.586477 +trainer/policy/normal/std Max 7.04195 +trainer/policy/normal/std Min 0.401893 +trainer/policy/normal/log_std Mean 1.0239 +trainer/policy/normal/log_std Std 0.262362 +trainer/policy/normal/log_std Max 1.95188 +trainer/policy/normal/log_std Min -0.91157 +eval/num steps total 204303 +eval/num paths total 205 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0703783 +eval/Actions Std 0.859717 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61935 +time/logging (s) 0.00376503 +time/sampling batch (s) 0.277031 +time/saving (s) 0.00341016 +time/training (s) 6.79466 +time/epoch (s) 9.69822 +time/total (s) 2173.34 +Epoch -796 +---------------------------------- --------------- +2022-05-10 13:47:00.105101 PDT | [1] Epoch -795 finished +---------------------------------- --------------- +epoch -795 +replay_buffer/size 999033 +trainer/num train calls 206000 +trainer/Policy Loss -19.2488 +trainer/Log Pis Mean 24.7602 +trainer/Log Pis Std 13.1423 +trainer/Log Pis Max 64.1782 +trainer/Log Pis Min -8.76715 +trainer/policy/mean Mean -0.0194915 +trainer/policy/mean Std 0.900308 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86759 +trainer/policy/normal/std Std 0.618557 +trainer/policy/normal/std Max 5.38067 +trainer/policy/normal/std Min 0.407414 +trainer/policy/normal/log_std Mean 1.02178 +trainer/policy/normal/log_std Std 0.2785 +trainer/policy/normal/log_std Max 1.68281 +trainer/policy/normal/log_std Min -0.897924 +eval/num steps total 205303 +eval/num paths total 206 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0599626 +eval/Actions Std 0.894666 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4768 +time/logging (s) 0.00369657 +time/sampling batch (s) 0.275391 +time/saving (s) 0.00340705 +time/training (s) 7.05258 +time/epoch (s) 9.81188 +time/total (s) 2183.15 +Epoch -795 +---------------------------------- --------------- +2022-05-10 13:47:10.259765 PDT | [1] Epoch -794 finished +---------------------------------- --------------- +epoch -794 +replay_buffer/size 999033 +trainer/num train calls 207000 +trainer/Policy Loss -18.2654 +trainer/Log Pis Mean 24.889 +trainer/Log Pis Std 13.6773 +trainer/Log Pis Max 70.3564 +trainer/Log Pis Min -6.47402 +trainer/policy/mean Mean -0.0552412 +trainer/policy/mean Std 0.903472 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.87945 +trainer/policy/normal/std Std 0.590198 +trainer/policy/normal/std Max 4.94716 +trainer/policy/normal/std Min 0.427288 +trainer/policy/normal/log_std Mean 1.02832 +trainer/policy/normal/log_std Std 0.269633 +trainer/policy/normal/log_std Max 1.59881 +trainer/policy/normal/log_std Min -0.850298 +eval/num steps total 206303 +eval/num paths total 207 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.289773 +eval/Actions Std 0.8383 +eval/Actions Max 0.999963 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57885 +time/logging (s) 0.00377134 +time/sampling batch (s) 0.525052 +time/saving (s) 0.00343235 +time/training (s) 7.02337 +time/epoch (s) 10.1345 +time/total (s) 2193.29 +Epoch -794 +---------------------------------- --------------- +2022-05-10 13:47:20.864381 PDT | [1] Epoch -793 finished +---------------------------------- --------------- +epoch -793 +replay_buffer/size 999033 +trainer/num train calls 208000 +trainer/Policy Loss -18.9578 +trainer/Log Pis Mean 24.6168 +trainer/Log Pis Std 12.8004 +trainer/Log Pis Max 65.0453 +trainer/Log Pis Min -6.10883 +trainer/policy/mean Mean -0.0504129 +trainer/policy/mean Std 0.905117 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.8743 +trainer/policy/normal/std Std 0.610591 +trainer/policy/normal/std Max 6.58728 +trainer/policy/normal/std Min 0.406525 +trainer/policy/normal/log_std Mean 1.0248 +trainer/policy/normal/log_std Std 0.27679 +trainer/policy/normal/log_std Max 1.88514 +trainer/policy/normal/log_std Min -0.900109 +eval/num steps total 207303 +eval/num paths total 208 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.25043 +eval/Actions Std 0.852374 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68634 +time/logging (s) 0.00478962 +time/sampling batch (s) 0.533341 +time/saving (s) 0.00527207 +time/training (s) 7.35484 +time/epoch (s) 10.5846 +time/total (s) 2203.88 +Epoch -793 +---------------------------------- --------------- +2022-05-10 13:47:31.114746 PDT | [1] Epoch -792 finished +---------------------------------- --------------- +epoch -792 +replay_buffer/size 999033 +trainer/num train calls 209000 +trainer/Policy Loss -18.6127 +trainer/Log Pis Mean 23.6991 +trainer/Log Pis Std 13.2297 +trainer/Log Pis Max 75.2402 +trainer/Log Pis Min -9.02267 +trainer/policy/mean Mean -0.0403533 +trainer/policy/mean Std 0.9055 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80081 +trainer/policy/normal/std Std 0.601361 +trainer/policy/normal/std Max 5.00194 +trainer/policy/normal/std Min 0.388817 +trainer/policy/normal/log_std Mean 0.997307 +trainer/policy/normal/log_std Std 0.285438 +trainer/policy/normal/log_std Max 1.60983 +trainer/policy/normal/log_std Min -0.944647 +eval/num steps total 208303 +eval/num paths total 209 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.427738 +eval/Actions Std 0.871431 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5957 +time/logging (s) 0.00409166 +time/sampling batch (s) 0.275523 +time/saving (s) 0.00394404 +time/training (s) 7.34988 +time/epoch (s) 10.2291 +time/total (s) 2214.11 +Epoch -792 +---------------------------------- --------------- +2022-05-10 13:47:42.360488 PDT | [1] Epoch -791 finished +---------------------------------- --------------- +epoch -791 +replay_buffer/size 999033 +trainer/num train calls 210000 +trainer/Policy Loss -19.4271 +trainer/Log Pis Mean 23.5968 +trainer/Log Pis Std 12.4213 +trainer/Log Pis Max 57.5852 +trainer/Log Pis Min -7.8535 +trainer/policy/mean Mean -0.0255406 +trainer/policy/mean Std 0.907349 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83052 +trainer/policy/normal/std Std 0.596826 +trainer/policy/normal/std Max 5.33038 +trainer/policy/normal/std Min 0.395857 +trainer/policy/normal/log_std Mean 1.00909 +trainer/policy/normal/log_std Std 0.280053 +trainer/policy/normal/log_std Max 1.67342 +trainer/policy/normal/log_std Min -0.926703 +eval/num steps total 209303 +eval/num paths total 210 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0455111 +eval/Actions Std 0.891218 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53506 +time/logging (s) 0.00395052 +time/sampling batch (s) 0.277544 +time/saving (s) 0.00360535 +time/training (s) 8.40435 +time/epoch (s) 11.2245 +time/total (s) 2225.34 +Epoch -791 +---------------------------------- --------------- +2022-05-10 13:47:53.742749 PDT | [1] Epoch -790 finished +---------------------------------- --------------- +epoch -790 +replay_buffer/size 999033 +trainer/num train calls 211000 +trainer/Policy Loss -18.2326 +trainer/Log Pis Mean 23.4579 +trainer/Log Pis Std 13.4388 +trainer/Log Pis Max 69.8154 +trainer/Log Pis Min -10.1049 +trainer/policy/mean Mean -0.0274442 +trainer/policy/mean Std 0.899634 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82997 +trainer/policy/normal/std Std 0.591185 +trainer/policy/normal/std Max 5.21853 +trainer/policy/normal/std Min 0.404703 +trainer/policy/normal/log_std Mean 1.0104 +trainer/policy/normal/log_std Std 0.270694 +trainer/policy/normal/log_std Max 1.65222 +trainer/policy/normal/log_std Min -0.904602 +eval/num steps total 210303 +eval/num paths total 211 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.148099 +eval/Actions Std 0.880729 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37879 +time/logging (s) 0.00433795 +time/sampling batch (s) 0.538057 +time/saving (s) 0.00405622 +time/training (s) 8.43592 +time/epoch (s) 11.3612 +time/total (s) 2236.7 +Epoch -790 +---------------------------------- --------------- +2022-05-10 13:48:04.204257 PDT | [1] Epoch -789 finished +---------------------------------- --------------- +epoch -789 +replay_buffer/size 999033 +trainer/num train calls 212000 +trainer/Policy Loss -20.0235 +trainer/Log Pis Mean 24.1816 +trainer/Log Pis Std 13.1406 +trainer/Log Pis Max 68.7326 +trainer/Log Pis Min -8.45581 +trainer/policy/mean Mean -0.0577389 +trainer/policy/mean Std 0.906265 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83426 +trainer/policy/normal/std Std 0.588241 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.342458 +trainer/policy/normal/log_std Mean 1.01253 +trainer/policy/normal/log_std Std 0.268456 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.07161 +eval/num steps total 211303 +eval/num paths total 212 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0112536 +eval/Actions Std 0.944939 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65119 +time/logging (s) 0.00420333 +time/sampling batch (s) 0.279164 +time/saving (s) 0.00368945 +time/training (s) 7.50174 +time/epoch (s) 10.44 +time/total (s) 2247.14 +Epoch -789 +---------------------------------- --------------- +2022-05-10 13:48:14.593459 PDT | [1] Epoch -788 finished +---------------------------------- --------------- +epoch -788 +replay_buffer/size 999033 +trainer/num train calls 213000 +trainer/Policy Loss -18.5124 +trainer/Log Pis Mean 23.4528 +trainer/Log Pis Std 12.9961 +trainer/Log Pis Max 79.1122 +trainer/Log Pis Min -12.4969 +trainer/policy/mean Mean -0.0283624 +trainer/policy/mean Std 0.903936 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82188 +trainer/policy/normal/std Std 0.575823 +trainer/policy/normal/std Max 4.82622 +trainer/policy/normal/std Min 0.372635 +trainer/policy/normal/log_std Mean 1.00866 +trainer/policy/normal/log_std Std 0.266494 +trainer/policy/normal/log_std Max 1.57406 +trainer/policy/normal/log_std Min -0.987155 +eval/num steps total 212303 +eval/num paths total 213 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00603413 +eval/Actions Std 0.925091 +eval/Actions Max 0.999999 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64648 +time/logging (s) 0.00438238 +time/sampling batch (s) 0.286713 +time/saving (s) 0.00348376 +time/training (s) 7.4271 +time/epoch (s) 10.3682 +time/total (s) 2257.51 +Epoch -788 +---------------------------------- --------------- +2022-05-10 13:48:25.890110 PDT | [1] Epoch -787 finished +---------------------------------- --------------- +epoch -787 +replay_buffer/size 999033 +trainer/num train calls 214000 +trainer/Policy Loss -20.2494 +trainer/Log Pis Mean 24.3983 +trainer/Log Pis Std 13.2237 +trainer/Log Pis Max 79.5903 +trainer/Log Pis Min -5.03719 +trainer/policy/mean Mean -0.0305982 +trainer/policy/mean Std 0.907027 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82786 +trainer/policy/normal/std Std 0.588243 +trainer/policy/normal/std Max 4.74496 +trainer/policy/normal/std Min 0.408517 +trainer/policy/normal/log_std Mean 1.00813 +trainer/policy/normal/log_std Std 0.282044 +trainer/policy/normal/log_std Max 1.55708 +trainer/policy/normal/log_std Min -0.895222 +eval/num steps total 213303 +eval/num paths total 214 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.284571 +eval/Actions Std 0.899111 +eval/Actions Max 0.999984 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61707 +time/logging (s) 0.00402141 +time/sampling batch (s) 0.280982 +time/saving (s) 0.00378212 +time/training (s) 8.36888 +time/epoch (s) 11.2747 +time/total (s) 2268.79 +Epoch -787 +---------------------------------- --------------- +2022-05-10 13:48:35.743169 PDT | [1] Epoch -786 finished +---------------------------------- --------------- +epoch -786 +replay_buffer/size 999033 +trainer/num train calls 215000 +trainer/Policy Loss -19.7857 +trainer/Log Pis Mean 24.3352 +trainer/Log Pis Std 12.9975 +trainer/Log Pis Max 64.6709 +trainer/Log Pis Min -11.0013 +trainer/policy/mean Mean -0.0534928 +trainer/policy/mean Std 0.899723 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80924 +trainer/policy/normal/std Std 0.600723 +trainer/policy/normal/std Max 5.02805 +trainer/policy/normal/std Min 0.391674 +trainer/policy/normal/log_std Mean 1.00054 +trainer/policy/normal/log_std Std 0.284392 +trainer/policy/normal/log_std Max 1.61503 +trainer/policy/normal/log_std Min -0.937326 +eval/num steps total 214303 +eval/num paths total 215 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.24324 +eval/Actions Std 0.844583 +eval/Actions Max 0.999951 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7342 +time/logging (s) 0.00375604 +time/sampling batch (s) 0.274648 +time/saving (s) 0.0036633 +time/training (s) 6.81566 +time/epoch (s) 9.83192 +time/total (s) 2278.63 +Epoch -786 +---------------------------------- --------------- +2022-05-10 13:48:45.461023 PDT | [1] Epoch -785 finished +---------------------------------- --------------- +epoch -785 +replay_buffer/size 999033 +trainer/num train calls 216000 +trainer/Policy Loss -20.0323 +trainer/Log Pis Mean 25.3508 +trainer/Log Pis Std 13.6722 +trainer/Log Pis Max 72.4331 +trainer/Log Pis Min -11.4544 +trainer/policy/mean Mean -0.014252 +trainer/policy/mean Std 0.909085 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.90521 +trainer/policy/normal/std Std 0.61061 +trainer/policy/normal/std Max 5.79956 +trainer/policy/normal/std Min 0.359936 +trainer/policy/normal/log_std Mean 1.03495 +trainer/policy/normal/log_std Std 0.281378 +trainer/policy/normal/log_std Max 1.75778 +trainer/policy/normal/log_std Min -1.02183 +eval/num steps total 215303 +eval/num paths total 216 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0518378 +eval/Actions Std 0.913667 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60916 +time/logging (s) 0.00377027 +time/sampling batch (s) 0.275111 +time/saving (s) 0.00343175 +time/training (s) 6.8058 +time/epoch (s) 9.69727 +time/total (s) 2288.33 +Epoch -785 +---------------------------------- --------------- +2022-05-10 13:48:55.168463 PDT | [1] Epoch -784 finished +---------------------------------- --------------- +epoch -784 +replay_buffer/size 999033 +trainer/num train calls 217000 +trainer/Policy Loss -21.0657 +trainer/Log Pis Mean 23.1643 +trainer/Log Pis Std 12.5456 +trainer/Log Pis Max 63.7298 +trainer/Log Pis Min -8.66071 +trainer/policy/mean Mean -0.0339867 +trainer/policy/mean Std 0.902985 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80139 +trainer/policy/normal/std Std 0.586115 +trainer/policy/normal/std Max 4.81402 +trainer/policy/normal/std Min 0.439708 +trainer/policy/normal/log_std Mean 0.999002 +trainer/policy/normal/log_std Std 0.279203 +trainer/policy/normal/log_std Max 1.57153 +trainer/policy/normal/log_std Min -0.821643 +eval/num steps total 216303 +eval/num paths total 217 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0440641 +eval/Actions Std 0.925018 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62005 +time/logging (s) 0.00368466 +time/sampling batch (s) 0.274844 +time/saving (s) 0.00340746 +time/training (s) 6.78495 +time/epoch (s) 9.68693 +time/total (s) 2298.02 +Epoch -784 +---------------------------------- --------------- +2022-05-10 13:49:06.063426 PDT | [1] Epoch -783 finished +---------------------------------- --------------- +epoch -783 +replay_buffer/size 999033 +trainer/num train calls 218000 +trainer/Policy Loss -19.7135 +trainer/Log Pis Mean 25.3092 +trainer/Log Pis Std 13.5174 +trainer/Log Pis Max 68.5076 +trainer/Log Pis Min -7.50134 +trainer/policy/mean Mean -0.0471288 +trainer/policy/mean Std 0.906 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87617 +trainer/policy/normal/std Std 0.594245 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.403685 +trainer/policy/normal/log_std Mean 1.02814 +trainer/policy/normal/log_std Std 0.261784 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.90712 +eval/num steps total 217303 +eval/num paths total 218 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10398 +eval/Actions Std 0.8802 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55378 +time/logging (s) 0.00371577 +time/sampling batch (s) 0.275087 +time/saving (s) 0.00342497 +time/training (s) 8.03847 +time/epoch (s) 10.8745 +time/total (s) 2308.89 +Epoch -783 +---------------------------------- --------------- +2022-05-10 13:49:16.308202 PDT | [1] Epoch -782 finished +---------------------------------- --------------- +epoch -782 +replay_buffer/size 999033 +trainer/num train calls 219000 +trainer/Policy Loss -19.344 +trainer/Log Pis Mean 23.9575 +trainer/Log Pis Std 13.3794 +trainer/Log Pis Max 74.0203 +trainer/Log Pis Min -7.59352 +trainer/policy/mean Mean -0.0370299 +trainer/policy/mean Std 0.905681 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80127 +trainer/policy/normal/std Std 0.578128 +trainer/policy/normal/std Max 5.64682 +trainer/policy/normal/std Min 0.377556 +trainer/policy/normal/log_std Mean 1.00032 +trainer/policy/normal/log_std Std 0.271706 +trainer/policy/normal/log_std Max 1.73109 +trainer/policy/normal/log_std Min -0.974037 +eval/num steps total 218303 +eval/num paths total 219 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0139964 +eval/Actions Std 0.908269 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5417 +time/logging (s) 0.00408376 +time/sampling batch (s) 0.525537 +time/saving (s) 0.00377476 +time/training (s) 7.14964 +time/epoch (s) 10.2247 +time/total (s) 2319.12 +Epoch -782 +---------------------------------- --------------- +2022-05-10 13:49:26.197398 PDT | [1] Epoch -781 finished +---------------------------------- --------------- +epoch -781 +replay_buffer/size 999033 +trainer/num train calls 220000 +trainer/Policy Loss -20.3823 +trainer/Log Pis Mean 25.1596 +trainer/Log Pis Std 12.7685 +trainer/Log Pis Max 63.6127 +trainer/Log Pis Min -6.30127 +trainer/policy/mean Mean -0.038725 +trainer/policy/mean Std 0.905912 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84238 +trainer/policy/normal/std Std 0.60198 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.399315 +trainer/policy/normal/log_std Mean 1.01299 +trainer/policy/normal/log_std Std 0.281671 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.918005 +eval/num steps total 219303 +eval/num paths total 220 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.257775 +eval/Actions Std 0.894664 +eval/Actions Max 0.999994 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70937 +time/logging (s) 0.00378779 +time/sampling batch (s) 0.524493 +time/saving (s) 0.00365585 +time/training (s) 6.62679 +time/epoch (s) 9.8681 +time/total (s) 2328.99 +Epoch -781 +---------------------------------- --------------- +2022-05-10 13:49:36.394594 PDT | [1] Epoch -780 finished +---------------------------------- --------------- +epoch -780 +replay_buffer/size 999033 +trainer/num train calls 221000 +trainer/Policy Loss -20.1103 +trainer/Log Pis Mean 24.5283 +trainer/Log Pis Std 13.4081 +trainer/Log Pis Max 72.3542 +trainer/Log Pis Min -7.57211 +trainer/policy/mean Mean -0.0401884 +trainer/policy/mean Std 0.909341 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86458 +trainer/policy/normal/std Std 0.609613 +trainer/policy/normal/std Max 5.24308 +trainer/policy/normal/std Min 0.315448 +trainer/policy/normal/log_std Mean 1.01882 +trainer/policy/normal/log_std Std 0.294358 +trainer/policy/normal/log_std Max 1.65691 +trainer/policy/normal/log_std Min -1.15376 +eval/num steps total 220303 +eval/num paths total 221 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.219315 +eval/Actions Std 0.958735 +eval/Actions Max 0.999972 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45585 +time/logging (s) 0.00376757 +time/sampling batch (s) 0.531613 +time/saving (s) 0.00341873 +time/training (s) 7.18151 +time/epoch (s) 10.1762 +time/total (s) 2339.17 +Epoch -780 +---------------------------------- --------------- +2022-05-10 13:49:46.616948 PDT | [1] Epoch -779 finished +---------------------------------- --------------- +epoch -779 +replay_buffer/size 999033 +trainer/num train calls 222000 +trainer/Policy Loss -19.0074 +trainer/Log Pis Mean 23.9024 +trainer/Log Pis Std 12.7305 +trainer/Log Pis Max 63.6949 +trainer/Log Pis Min -7.51099 +trainer/policy/mean Mean -0.0375931 +trainer/policy/mean Std 0.904422 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87672 +trainer/policy/normal/std Std 0.587128 +trainer/policy/normal/std Max 5.51367 +trainer/policy/normal/std Min 0.397767 +trainer/policy/normal/log_std Mean 1.0277 +trainer/policy/normal/log_std Std 0.268147 +trainer/policy/normal/log_std Max 1.70723 +trainer/policy/normal/log_std Min -0.92189 +eval/num steps total 221303 +eval/num paths total 222 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.336227 +eval/Actions Std 0.793242 +eval/Actions Max 0.999999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47842 +time/logging (s) 0.00367985 +time/sampling batch (s) 0.281104 +time/saving (s) 0.00351244 +time/training (s) 7.43458 +time/epoch (s) 10.2013 +time/total (s) 2349.38 +Epoch -779 +---------------------------------- --------------- +2022-05-10 13:49:56.924816 PDT | [1] Epoch -778 finished +---------------------------------- --------------- +epoch -778 +replay_buffer/size 999033 +trainer/num train calls 223000 +trainer/Policy Loss -19.2815 +trainer/Log Pis Mean 24.1953 +trainer/Log Pis Std 13.2454 +trainer/Log Pis Max 71.1809 +trainer/Log Pis Min -4.84971 +trainer/policy/mean Mean -0.027781 +trainer/policy/mean Std 0.90905 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83049 +trainer/policy/normal/std Std 0.578887 +trainer/policy/normal/std Max 5.00417 +trainer/policy/normal/std Min 0.468325 +trainer/policy/normal/log_std Mean 1.01281 +trainer/policy/normal/log_std Std 0.257173 +trainer/policy/normal/log_std Max 1.61027 +trainer/policy/normal/log_std Min -0.758592 +eval/num steps total 222303 +eval/num paths total 223 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.141426 +eval/Actions Std 0.879653 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60859 +time/logging (s) 0.00380981 +time/sampling batch (s) 0.282368 +time/saving (s) 0.00354151 +time/training (s) 7.38839 +time/epoch (s) 10.2867 +time/total (s) 2359.67 +Epoch -778 +---------------------------------- --------------- +2022-05-10 13:50:07.618405 PDT | [1] Epoch -777 finished +---------------------------------- --------------- +epoch -777 +replay_buffer/size 999033 +trainer/num train calls 224000 +trainer/Policy Loss -18.7938 +trainer/Log Pis Mean 23.8916 +trainer/Log Pis Std 12.9316 +trainer/Log Pis Max 58.0654 +trainer/Log Pis Min -4.61741 +trainer/policy/mean Mean -0.0180342 +trainer/policy/mean Std 0.905229 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83822 +trainer/policy/normal/std Std 0.603546 +trainer/policy/normal/std Max 5.07593 +trainer/policy/normal/std Min 0.401056 +trainer/policy/normal/log_std Mean 1.01173 +trainer/policy/normal/log_std Std 0.278971 +trainer/policy/normal/log_std Max 1.62451 +trainer/policy/normal/log_std Min -0.913655 +eval/num steps total 223303 +eval/num paths total 224 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0464571 +eval/Actions Std 0.91109 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57963 +time/logging (s) 0.00394627 +time/sampling batch (s) 0.779512 +time/saving (s) 0.00350774 +time/training (s) 7.30613 +time/epoch (s) 10.6727 +time/total (s) 2370.34 +Epoch -777 +---------------------------------- --------------- +2022-05-10 13:50:18.117828 PDT | [1] Epoch -776 finished +---------------------------------- --------------- +epoch -776 +replay_buffer/size 999033 +trainer/num train calls 225000 +trainer/Policy Loss -19.2343 +trainer/Log Pis Mean 24.1602 +trainer/Log Pis Std 13.2794 +trainer/Log Pis Max 64.1655 +trainer/Log Pis Min -11.0796 +trainer/policy/mean Mean -0.0430587 +trainer/policy/mean Std 0.90781 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86329 +trainer/policy/normal/std Std 0.597421 +trainer/policy/normal/std Max 5.37884 +trainer/policy/normal/std Min 0.405146 +trainer/policy/normal/log_std Mean 1.02182 +trainer/policy/normal/log_std Std 0.273231 +trainer/policy/normal/log_std Max 1.68247 +trainer/policy/normal/log_std Min -0.903508 +eval/num steps total 224303 +eval/num paths total 225 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00938882 +eval/Actions Std 0.972426 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60275 +time/logging (s) 0.00397939 +time/sampling batch (s) 0.528382 +time/saving (s) 0.00390605 +time/training (s) 7.3397 +time/epoch (s) 10.4787 +time/total (s) 2380.82 +Epoch -776 +---------------------------------- --------------- +2022-05-10 13:50:28.425894 PDT | [1] Epoch -775 finished +---------------------------------- --------------- +epoch -775 +replay_buffer/size 999033 +trainer/num train calls 226000 +trainer/Policy Loss -18.4858 +trainer/Log Pis Mean 24.2656 +trainer/Log Pis Std 13.1322 +trainer/Log Pis Max 70.0369 +trainer/Log Pis Min -10.4781 +trainer/policy/mean Mean -0.0530455 +trainer/policy/mean Std 0.901469 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8607 +trainer/policy/normal/std Std 0.591503 +trainer/policy/normal/std Max 5.36034 +trainer/policy/normal/std Min 0.404565 +trainer/policy/normal/log_std Mean 1.02117 +trainer/policy/normal/log_std Std 0.27296 +trainer/policy/normal/log_std Max 1.67903 +trainer/policy/normal/log_std Min -0.904942 +eval/num steps total 225303 +eval/num paths total 226 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.343221 +eval/Actions Std 0.782049 +eval/Actions Max 0.999997 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62804 +time/logging (s) 0.00389669 +time/sampling batch (s) 0.775415 +time/saving (s) 0.00348643 +time/training (s) 6.87619 +time/epoch (s) 10.287 +time/total (s) 2391.11 +Epoch -775 +---------------------------------- --------------- +2022-05-10 13:50:37.775685 PDT | [1] Epoch -774 finished +---------------------------------- --------------- +epoch -774 +replay_buffer/size 999033 +trainer/num train calls 227000 +trainer/Policy Loss -19.2949 +trainer/Log Pis Mean 25.2734 +trainer/Log Pis Std 13.4248 +trainer/Log Pis Max 71.9818 +trainer/Log Pis Min -5.9532 +trainer/policy/mean Mean -0.0440068 +trainer/policy/mean Std 0.909586 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.94522 +trainer/policy/normal/std Std 0.631194 +trainer/policy/normal/std Max 6.751 +trainer/policy/normal/std Min 0.44484 +trainer/policy/normal/log_std Mean 1.0491 +trainer/policy/normal/log_std Std 0.276022 +trainer/policy/normal/log_std Max 1.90969 +trainer/policy/normal/log_std Min -0.81004 +eval/num steps total 226303 +eval/num paths total 227 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.124349 +eval/Actions Std 0.924624 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49983 +time/logging (s) 0.0037754 +time/sampling batch (s) 0.274983 +time/saving (s) 0.00342914 +time/training (s) 6.54702 +time/epoch (s) 9.32904 +time/total (s) 2400.45 +Epoch -774 +---------------------------------- --------------- +2022-05-10 13:50:48.293018 PDT | [1] Epoch -773 finished +---------------------------------- --------------- +epoch -773 +replay_buffer/size 999033 +trainer/num train calls 228000 +trainer/Policy Loss -19.3262 +trainer/Log Pis Mean 23.7984 +trainer/Log Pis Std 12.8654 +trainer/Log Pis Max 65.4894 +trainer/Log Pis Min -8.86558 +trainer/policy/mean Mean -0.0232969 +trainer/policy/mean Std 0.905727 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.89855 +trainer/policy/normal/std Std 0.613072 +trainer/policy/normal/std Max 5.39164 +trainer/policy/normal/std Min 0.377762 +trainer/policy/normal/log_std Mean 1.0339 +trainer/policy/normal/log_std Std 0.272431 +trainer/policy/normal/log_std Max 1.68485 +trainer/policy/normal/log_std Min -0.97349 +eval/num steps total 227303 +eval/num paths total 228 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.349731 +eval/Actions Std 0.764584 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6269 +time/logging (s) 0.00371259 +time/sampling batch (s) 0.525738 +time/saving (s) 0.00343405 +time/training (s) 7.33694 +time/epoch (s) 10.4967 +time/total (s) 2410.95 +Epoch -773 +---------------------------------- --------------- +2022-05-10 13:50:58.545137 PDT | [1] Epoch -772 finished +---------------------------------- --------------- +epoch -772 +replay_buffer/size 999033 +trainer/num train calls 229000 +trainer/Policy Loss -20.2819 +trainer/Log Pis Mean 24.4186 +trainer/Log Pis Std 13.2236 +trainer/Log Pis Max 74.4011 +trainer/Log Pis Min -9.86921 +trainer/policy/mean Mean -0.0102774 +trainer/policy/mean Std 0.906813 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.9405 +trainer/policy/normal/std Std 0.625041 +trainer/policy/normal/std Max 6.52319 +trainer/policy/normal/std Min 0.419998 +trainer/policy/normal/log_std Mean 1.04763 +trainer/policy/normal/log_std Std 0.27621 +trainer/policy/normal/log_std Max 1.87536 +trainer/policy/normal/log_std Min -0.867505 +eval/num steps total 228303 +eval/num paths total 229 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114557 +eval/Actions Std 0.917491 +eval/Actions Max 0.999974 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64904 +time/logging (s) 0.00383111 +time/sampling batch (s) 0.524476 +time/saving (s) 0.00347467 +time/training (s) 7.05095 +time/epoch (s) 10.2318 +time/total (s) 2421.18 +Epoch -772 +---------------------------------- --------------- +2022-05-10 13:51:08.571621 PDT | [1] Epoch -771 finished +---------------------------------- --------------- +epoch -771 +replay_buffer/size 999033 +trainer/num train calls 230000 +trainer/Policy Loss -18.9917 +trainer/Log Pis Mean 24.9283 +trainer/Log Pis Std 13.3256 +trainer/Log Pis Max 69.895 +trainer/Log Pis Min -8.84604 +trainer/policy/mean Mean -0.0352009 +trainer/policy/mean Std 0.902122 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85451 +trainer/policy/normal/std Std 0.605066 +trainer/policy/normal/std Max 7.29211 +trainer/policy/normal/std Min 0.357081 +trainer/policy/normal/log_std Mean 1.01845 +trainer/policy/normal/log_std Std 0.273663 +trainer/policy/normal/log_std Max 1.98679 +trainer/policy/normal/log_std Min -1.02979 +eval/num steps total 229303 +eval/num paths total 230 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118331 +eval/Actions Std 0.902601 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39409 +time/logging (s) 0.00425188 +time/sampling batch (s) 0.275278 +time/saving (s) 0.00396643 +time/training (s) 7.32879 +time/epoch (s) 10.0064 +time/total (s) 2431.19 +Epoch -771 +---------------------------------- --------------- +2022-05-10 13:51:19.715942 PDT | [1] Epoch -770 finished +---------------------------------- --------------- +epoch -770 +replay_buffer/size 999033 +trainer/num train calls 231000 +trainer/Policy Loss -20.3592 +trainer/Log Pis Mean 25.5333 +trainer/Log Pis Std 13.1724 +trainer/Log Pis Max 62.5081 +trainer/Log Pis Min -4.46044 +trainer/policy/mean Mean -0.0358715 +trainer/policy/mean Std 0.908131 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86815 +trainer/policy/normal/std Std 0.606105 +trainer/policy/normal/std Max 5.68727 +trainer/policy/normal/std Min 0.377293 +trainer/policy/normal/log_std Mean 1.02266 +trainer/policy/normal/log_std Std 0.276912 +trainer/policy/normal/log_std Max 1.73823 +trainer/policy/normal/log_std Min -0.974733 +eval/num steps total 230303 +eval/num paths total 231 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.257153 +eval/Actions Std 0.877854 +eval/Actions Max 0.999984 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4158 +time/logging (s) 0.003725 +time/sampling batch (s) 0.527679 +time/saving (s) 0.00354947 +time/training (s) 8.1719 +time/epoch (s) 11.1227 +time/total (s) 2442.32 +Epoch -770 +---------------------------------- --------------- +2022-05-10 13:51:28.949099 PDT | [1] Epoch -769 finished +---------------------------------- --------------- +epoch -769 +replay_buffer/size 999033 +trainer/num train calls 232000 +trainer/Policy Loss -19.2507 +trainer/Log Pis Mean 23.987 +trainer/Log Pis Std 13.0542 +trainer/Log Pis Max 67.6787 +trainer/Log Pis Min -9.46073 +trainer/policy/mean Mean -0.0139342 +trainer/policy/mean Std 0.906314 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.78342 +trainer/policy/normal/std Std 0.614348 +trainer/policy/normal/std Max 5.42621 +trainer/policy/normal/std Min 0.408604 +trainer/policy/normal/log_std Mean 0.988267 +trainer/policy/normal/log_std Std 0.299754 +trainer/policy/normal/log_std Max 1.69124 +trainer/policy/normal/log_std Min -0.89501 +eval/num steps total 231303 +eval/num paths total 232 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0700651 +eval/Actions Std 0.882795 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62182 +time/logging (s) 0.00371631 +time/sampling batch (s) 0.274717 +time/saving (s) 0.00344715 +time/training (s) 6.30884 +time/epoch (s) 9.21254 +time/total (s) 2451.53 +Epoch -769 +---------------------------------- --------------- +2022-05-10 13:51:38.329598 PDT | [1] Epoch -768 finished +---------------------------------- --------------- +epoch -768 +replay_buffer/size 999033 +trainer/num train calls 233000 +trainer/Policy Loss -18.7774 +trainer/Log Pis Mean 23.7296 +trainer/Log Pis Std 13.0663 +trainer/Log Pis Max 62.8672 +trainer/Log Pis Min -6.05274 +trainer/policy/mean Mean -0.0394448 +trainer/policy/mean Std 0.909424 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84289 +trainer/policy/normal/std Std 0.588493 +trainer/policy/normal/std Max 5.72786 +trainer/policy/normal/std Min 0.459851 +trainer/policy/normal/log_std Mean 1.01492 +trainer/policy/normal/log_std Std 0.272675 +trainer/policy/normal/log_std Max 1.74534 +trainer/policy/normal/log_std Min -0.776853 +eval/num steps total 232303 +eval/num paths total 233 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.064877 +eval/Actions Std 0.911192 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38659 +time/logging (s) 0.00374623 +time/sampling batch (s) 0.27817 +time/saving (s) 0.00357632 +time/training (s) 6.68767 +time/epoch (s) 9.35975 +time/total (s) 2460.89 +Epoch -768 +---------------------------------- --------------- +2022-05-10 13:51:48.079297 PDT | [1] Epoch -767 finished +---------------------------------- --------------- +epoch -767 +replay_buffer/size 999033 +trainer/num train calls 234000 +trainer/Policy Loss -18.687 +trainer/Log Pis Mean 23.7751 +trainer/Log Pis Std 12.3973 +trainer/Log Pis Max 61.2637 +trainer/Log Pis Min -7.88568 +trainer/policy/mean Mean -0.0464714 +trainer/policy/mean Std 0.904582 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8568 +trainer/policy/normal/std Std 0.591814 +trainer/policy/normal/std Max 5.94372 +trainer/policy/normal/std Min 0.356605 +trainer/policy/normal/log_std Mean 1.02 +trainer/policy/normal/log_std Std 0.271425 +trainer/policy/normal/log_std Max 1.78233 +trainer/policy/normal/log_std Min -1.03113 +eval/num steps total 233303 +eval/num paths total 234 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.313562 +eval/Actions Std 0.868538 +eval/Actions Max 0.999995 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49527 +time/logging (s) 0.00370199 +time/sampling batch (s) 0.27908 +time/saving (s) 0.00356903 +time/training (s) 6.94706 +time/epoch (s) 9.72868 +time/total (s) 2470.63 +Epoch -767 +---------------------------------- --------------- +2022-05-10 13:51:57.506347 PDT | [1] Epoch -766 finished +---------------------------------- --------------- +epoch -766 +replay_buffer/size 999033 +trainer/num train calls 235000 +trainer/Policy Loss -19.4061 +trainer/Log Pis Mean 24.8921 +trainer/Log Pis Std 13.4112 +trainer/Log Pis Max 81.35 +trainer/Log Pis Min -6.76777 +trainer/policy/mean Mean -0.0266401 +trainer/policy/mean Std 0.901164 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85054 +trainer/policy/normal/std Std 0.594766 +trainer/policy/normal/std Max 5.95236 +trainer/policy/normal/std Min 0.407145 +trainer/policy/normal/log_std Mean 1.01813 +trainer/policy/normal/log_std Std 0.268086 +trainer/policy/normal/log_std Max 1.78379 +trainer/policy/normal/log_std Min -0.898585 +eval/num steps total 234303 +eval/num paths total 235 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.342367 +eval/Actions Std 0.876961 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88117 +time/logging (s) 0.004173 +time/sampling batch (s) 0.274721 +time/saving (s) 0.00402377 +time/training (s) 6.24266 +time/epoch (s) 9.40675 +time/total (s) 2480.04 +Epoch -766 +---------------------------------- --------------- +2022-05-10 13:52:07.193070 PDT | [1] Epoch -765 finished +---------------------------------- --------------- +epoch -765 +replay_buffer/size 999033 +trainer/num train calls 236000 +trainer/Policy Loss -20.7748 +trainer/Log Pis Mean 24.4377 +trainer/Log Pis Std 13.135 +trainer/Log Pis Max 60.2775 +trainer/Log Pis Min -5.50188 +trainer/policy/mean Mean -0.0261065 +trainer/policy/mean Std 0.904404 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81302 +trainer/policy/normal/std Std 0.57345 +trainer/policy/normal/std Max 4.87062 +trainer/policy/normal/std Min 0.389358 +trainer/policy/normal/log_std Mean 1.00547 +trainer/policy/normal/log_std Std 0.266525 +trainer/policy/normal/log_std Max 1.58322 +trainer/policy/normal/log_std Min -0.943257 +eval/num steps total 235303 +eval/num paths total 236 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.083446 +eval/Actions Std 0.903729 +eval/Actions Max 0.999988 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74632 +time/logging (s) 0.0038064 +time/sampling batch (s) 0.27856 +time/saving (s) 0.00378371 +time/training (s) 6.63245 +time/epoch (s) 9.66491 +time/total (s) 2489.7 +Epoch -765 +---------------------------------- --------------- +2022-05-10 13:52:16.215210 PDT | [1] Epoch -764 finished +---------------------------------- --------------- +epoch -764 +replay_buffer/size 999033 +trainer/num train calls 237000 +trainer/Policy Loss -19.3204 +trainer/Log Pis Mean 24.6831 +trainer/Log Pis Std 14.0788 +trainer/Log Pis Max 80.8 +trainer/Log Pis Min -10.0046 +trainer/policy/mean Mean -0.0293508 +trainer/policy/mean Std 0.906716 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.861 +trainer/policy/normal/std Std 0.615719 +trainer/policy/normal/std Max 5.3334 +trainer/policy/normal/std Min 0.415236 +trainer/policy/normal/log_std Mean 1.01868 +trainer/policy/normal/log_std Std 0.284747 +trainer/policy/normal/log_std Max 1.67399 +trainer/policy/normal/log_std Min -0.878908 +eval/num steps total 236303 +eval/num paths total 237 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00303397 +eval/Actions Std 0.906407 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42623 +time/logging (s) 0.00366409 +time/sampling batch (s) 0.28037 +time/saving (s) 0.00351884 +time/training (s) 6.2871 +time/epoch (s) 9.00089 +time/total (s) 2498.71 +Epoch -764 +---------------------------------- --------------- +2022-05-10 13:52:25.922693 PDT | [1] Epoch -763 finished +---------------------------------- --------------- +epoch -763 +replay_buffer/size 999033 +trainer/num train calls 238000 +trainer/Policy Loss -19.8273 +trainer/Log Pis Mean 24.0745 +trainer/Log Pis Std 13.0864 +trainer/Log Pis Max 82.7799 +trainer/Log Pis Min -7.62157 +trainer/policy/mean Mean -0.0485769 +trainer/policy/mean Std 0.901715 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86617 +trainer/policy/normal/std Std 0.603997 +trainer/policy/normal/std Max 5.69744 +trainer/policy/normal/std Min 0.449697 +trainer/policy/normal/log_std Mean 1.02351 +trainer/policy/normal/log_std Std 0.266804 +trainer/policy/normal/log_std Max 1.74002 +trainer/policy/normal/log_std Min -0.799182 +eval/num steps total 237303 +eval/num paths total 238 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108832 +eval/Actions Std 0.90713 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68814 +time/logging (s) 0.00373743 +time/sampling batch (s) 0.531769 +time/saving (s) 0.0035258 +time/training (s) 6.45922 +time/epoch (s) 9.68639 +time/total (s) 2508.4 +Epoch -763 +---------------------------------- --------------- +2022-05-10 13:52:36.293912 PDT | [1] Epoch -762 finished +---------------------------------- --------------- +epoch -762 +replay_buffer/size 999033 +trainer/num train calls 239000 +trainer/Policy Loss -20.0309 +trainer/Log Pis Mean 24.2958 +trainer/Log Pis Std 12.8908 +trainer/Log Pis Max 70.0019 +trainer/Log Pis Min -5.62247 +trainer/policy/mean Mean -0.0399616 +trainer/policy/mean Std 0.908524 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86321 +trainer/policy/normal/std Std 0.612044 +trainer/policy/normal/std Max 5.44103 +trainer/policy/normal/std Min 0.300432 +trainer/policy/normal/log_std Mean 1.01945 +trainer/policy/normal/log_std Std 0.286461 +trainer/policy/normal/log_std Max 1.69397 +trainer/policy/normal/log_std Min -1.20253 +eval/num steps total 238303 +eval/num paths total 239 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0492989 +eval/Actions Std 0.899728 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63853 +time/logging (s) 0.00369499 +time/sampling batch (s) 0.283347 +time/saving (s) 0.00345503 +time/training (s) 7.42085 +time/epoch (s) 10.3499 +time/total (s) 2518.75 +Epoch -762 +---------------------------------- --------------- +2022-05-10 13:52:46.162748 PDT | [1] Epoch -761 finished +---------------------------------- --------------- +epoch -761 +replay_buffer/size 999033 +trainer/num train calls 240000 +trainer/Policy Loss -20.7099 +trainer/Log Pis Mean 23.0621 +trainer/Log Pis Std 13.3376 +trainer/Log Pis Max 71.6228 +trainer/Log Pis Min -5.47651 +trainer/policy/mean Mean -0.026289 +trainer/policy/mean Std 0.9019 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79817 +trainer/policy/normal/std Std 0.577557 +trainer/policy/normal/std Max 5.27846 +trainer/policy/normal/std Min 0.336084 +trainer/policy/normal/log_std Mean 0.999231 +trainer/policy/normal/log_std Std 0.272209 +trainer/policy/normal/log_std Max 1.66363 +trainer/policy/normal/log_std Min -1.09039 +eval/num steps total 239303 +eval/num paths total 240 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0870582 +eval/Actions Std 0.896111 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56334 +time/logging (s) 0.00373199 +time/sampling batch (s) 0.280961 +time/saving (s) 0.0034699 +time/training (s) 6.99632 +time/epoch (s) 9.84782 +time/total (s) 2528.6 +Epoch -761 +---------------------------------- --------------- +2022-05-10 13:52:56.568144 PDT | [1] Epoch -760 finished +---------------------------------- --------------- +epoch -760 +replay_buffer/size 999033 +trainer/num train calls 241000 +trainer/Policy Loss -19.5222 +trainer/Log Pis Mean 24.9841 +trainer/Log Pis Std 13.0522 +trainer/Log Pis Max 76.8169 +trainer/Log Pis Min -7.80845 +trainer/policy/mean Mean -0.026428 +trainer/policy/mean Std 0.910341 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82176 +trainer/policy/normal/std Std 0.591935 +trainer/policy/normal/std Max 5.3742 +trainer/policy/normal/std Min 0.286268 +trainer/policy/normal/log_std Mean 1.00585 +trainer/policy/normal/log_std Std 0.282817 +trainer/policy/normal/log_std Max 1.68161 +trainer/policy/normal/log_std Min -1.25083 +eval/num steps total 240303 +eval/num paths total 241 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.191442 +eval/Actions Std 0.89363 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69523 +time/logging (s) 0.00397968 +time/sampling batch (s) 0.531334 +time/saving (s) 0.00387025 +time/training (s) 7.15013 +time/epoch (s) 10.3845 +time/total (s) 2538.99 +Epoch -760 +---------------------------------- --------------- +2022-05-10 13:53:06.282506 PDT | [1] Epoch -759 finished +---------------------------------- --------------- +epoch -759 +replay_buffer/size 999033 +trainer/num train calls 242000 +trainer/Policy Loss -18.7959 +trainer/Log Pis Mean 23.8806 +trainer/Log Pis Std 13.2269 +trainer/Log Pis Max 73.0166 +trainer/Log Pis Min -8.98049 +trainer/policy/mean Mean -0.0231497 +trainer/policy/mean Std 0.904314 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80534 +trainer/policy/normal/std Std 0.590735 +trainer/policy/normal/std Max 5.2261 +trainer/policy/normal/std Min 0.302963 +trainer/policy/normal/log_std Mean 0.99968 +trainer/policy/normal/log_std Std 0.283905 +trainer/policy/normal/log_std Max 1.65367 +trainer/policy/normal/log_std Min -1.19415 +eval/num steps total 241303 +eval/num paths total 242 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0796275 +eval/Actions Std 0.888026 +eval/Actions Max 0.999999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57129 +time/logging (s) 0.0038419 +time/sampling batch (s) 0.281868 +time/saving (s) 0.00360716 +time/training (s) 6.83228 +time/epoch (s) 9.69289 +time/total (s) 2548.68 +Epoch -759 +---------------------------------- --------------- +2022-05-10 13:53:16.124442 PDT | [1] Epoch -758 finished +---------------------------------- --------------- +epoch -758 +replay_buffer/size 999033 +trainer/num train calls 243000 +trainer/Policy Loss -20.6779 +trainer/Log Pis Mean 23.8439 +trainer/Log Pis Std 13.9665 +trainer/Log Pis Max 75.9561 +trainer/Log Pis Min -7.24296 +trainer/policy/mean Mean -0.0293747 +trainer/policy/mean Std 0.901359 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87598 +trainer/policy/normal/std Std 0.604456 +trainer/policy/normal/std Max 5.64342 +trainer/policy/normal/std Min 0.345651 +trainer/policy/normal/log_std Mean 1.0255 +trainer/policy/normal/log_std Std 0.277355 +trainer/policy/normal/log_std Max 1.73049 +trainer/policy/normal/log_std Min -1.06233 +eval/num steps total 242303 +eval/num paths total 243 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00397261 +eval/Actions Std 0.959656 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77388 +time/logging (s) 0.00381405 +time/sampling batch (s) 0.532497 +time/saving (s) 0.00348089 +time/training (s) 6.50697 +time/epoch (s) 9.82064 +time/total (s) 2558.51 +Epoch -758 +---------------------------------- --------------- +2022-05-10 13:53:26.819429 PDT | [1] Epoch -757 finished +---------------------------------- --------------- +epoch -757 +replay_buffer/size 999033 +trainer/num train calls 244000 +trainer/Policy Loss -20.5916 +trainer/Log Pis Mean 23.6472 +trainer/Log Pis Std 12.6659 +trainer/Log Pis Max 69.834 +trainer/Log Pis Min -4.76904 +trainer/policy/mean Mean -0.0262717 +trainer/policy/mean Std 0.904979 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8405 +trainer/policy/normal/std Std 0.608599 +trainer/policy/normal/std Max 5.47123 +trainer/policy/normal/std Min 0.379843 +trainer/policy/normal/log_std Mean 1.01141 +trainer/policy/normal/log_std Std 0.286255 +trainer/policy/normal/log_std Max 1.6995 +trainer/policy/normal/log_std Min -0.967998 +eval/num steps total 243303 +eval/num paths total 244 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0581687 +eval/Actions Std 0.910354 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66059 +time/logging (s) 0.00381605 +time/sampling batch (s) 0.281565 +time/saving (s) 0.00365395 +time/training (s) 7.72389 +time/epoch (s) 10.6735 +time/total (s) 2569.18 +Epoch -757 +---------------------------------- --------------- +2022-05-10 13:53:37.231776 PDT | [1] Epoch -756 finished +---------------------------------- --------------- +epoch -756 +replay_buffer/size 999033 +trainer/num train calls 245000 +trainer/Policy Loss -19.2284 +trainer/Log Pis Mean 24.7271 +trainer/Log Pis Std 12.7954 +trainer/Log Pis Max 55.2648 +trainer/Log Pis Min -7.34295 +trainer/policy/mean Mean -0.0569959 +trainer/policy/mean Std 0.901122 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85547 +trainer/policy/normal/std Std 0.620675 +trainer/policy/normal/std Max 5.98833 +trainer/policy/normal/std Min 0.34339 +trainer/policy/normal/log_std Mean 1.01631 +trainer/policy/normal/log_std Std 0.287448 +trainer/policy/normal/log_std Max 1.78981 +trainer/policy/normal/log_std Min -1.06889 +eval/num steps total 244303 +eval/num paths total 245 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.13476 +eval/Actions Std 0.881118 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5794 +time/logging (s) 0.00368633 +time/sampling batch (s) 0.281951 +time/saving (s) 0.00342195 +time/training (s) 7.52238 +time/epoch (s) 10.3908 +time/total (s) 2579.58 +Epoch -756 +---------------------------------- --------------- +2022-05-10 13:53:46.878257 PDT | [1] Epoch -755 finished +---------------------------------- --------------- +epoch -755 +replay_buffer/size 999033 +trainer/num train calls 246000 +trainer/Policy Loss -18.7585 +trainer/Log Pis Mean 24.7473 +trainer/Log Pis Std 12.7874 +trainer/Log Pis Max 61.8193 +trainer/Log Pis Min -5.39599 +trainer/policy/mean Mean -0.0375137 +trainer/policy/mean Std 0.906311 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.89167 +trainer/policy/normal/std Std 0.631142 +trainer/policy/normal/std Max 5.56043 +trainer/policy/normal/std Min 0.402696 +trainer/policy/normal/log_std Mean 1.0271 +trainer/policy/normal/log_std Std 0.29784 +trainer/policy/normal/log_std Max 1.71568 +trainer/policy/normal/log_std Min -0.909573 +eval/num steps total 245303 +eval/num paths total 246 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0931725 +eval/Actions Std 0.914005 +eval/Actions Max 0.999991 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53442 +time/logging (s) 0.00415011 +time/sampling batch (s) 0.282148 +time/saving (s) 0.00405652 +time/training (s) 6.80109 +time/epoch (s) 9.62586 +time/total (s) 2589.21 +Epoch -755 +---------------------------------- --------------- +2022-05-10 13:53:56.843396 PDT | [1] Epoch -754 finished +---------------------------------- --------------- +epoch -754 +replay_buffer/size 999033 +trainer/num train calls 247000 +trainer/Policy Loss -18.9849 +trainer/Log Pis Mean 24.6492 +trainer/Log Pis Std 12.8354 +trainer/Log Pis Max 72.4929 +trainer/Log Pis Min -8.05012 +trainer/policy/mean Mean -0.039735 +trainer/policy/mean Std 0.908782 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81664 +trainer/policy/normal/std Std 0.590541 +trainer/policy/normal/std Max 5.37659 +trainer/policy/normal/std Min 0.365049 +trainer/policy/normal/log_std Mean 1.00471 +trainer/policy/normal/log_std Std 0.27737 +trainer/policy/normal/log_std Max 1.68205 +trainer/policy/normal/log_std Min -1.00772 +eval/num steps total 246303 +eval/num paths total 247 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0086921 +eval/Actions Std 0.906876 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63775 +time/logging (s) 0.00384694 +time/sampling batch (s) 0.280886 +time/saving (s) 0.00365734 +time/training (s) 7.0173 +time/epoch (s) 9.94344 +time/total (s) 2599.15 +Epoch -754 +---------------------------------- --------------- +2022-05-10 13:54:06.402957 PDT | [1] Epoch -753 finished +---------------------------------- --------------- +epoch -753 +replay_buffer/size 999033 +trainer/num train calls 248000 +trainer/Policy Loss -21.3213 +trainer/Log Pis Mean 24.5847 +trainer/Log Pis Std 13.2801 +trainer/Log Pis Max 75.2738 +trainer/Log Pis Min -9.44383 +trainer/policy/mean Mean -0.0297887 +trainer/policy/mean Std 0.905935 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81642 +trainer/policy/normal/std Std 0.616706 +trainer/policy/normal/std Max 5.50755 +trainer/policy/normal/std Min 0.444127 +trainer/policy/normal/log_std Mean 1.00257 +trainer/policy/normal/log_std Std 0.284446 +trainer/policy/normal/log_std Max 1.70612 +trainer/policy/normal/log_std Min -0.811646 +eval/num steps total 247303 +eval/num paths total 248 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.190345 +eval/Actions Std 0.924373 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71804 +time/logging (s) 0.00372894 +time/sampling batch (s) 0.529836 +time/saving (s) 0.00343842 +time/training (s) 6.2834 +time/epoch (s) 9.53844 +time/total (s) 2608.7 +Epoch -753 +---------------------------------- --------------- +2022-05-10 13:54:16.550528 PDT | [1] Epoch -752 finished +---------------------------------- --------------- +epoch -752 +replay_buffer/size 999033 +trainer/num train calls 249000 +trainer/Policy Loss -18.9461 +trainer/Log Pis Mean 24.6986 +trainer/Log Pis Std 13.7879 +trainer/Log Pis Max 76.1408 +trainer/Log Pis Min -9.68598 +trainer/policy/mean Mean -0.0301982 +trainer/policy/mean Std 0.909274 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84291 +trainer/policy/normal/std Std 0.614409 +trainer/policy/normal/std Max 6.80999 +trainer/policy/normal/std Min 0.439124 +trainer/policy/normal/log_std Mean 1.01263 +trainer/policy/normal/log_std Std 0.281967 +trainer/policy/normal/log_std Max 1.91839 +trainer/policy/normal/log_std Min -0.822973 +eval/num steps total 248303 +eval/num paths total 249 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124788 +eval/Actions Std 0.880731 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71747 +time/logging (s) 0.00378673 +time/sampling batch (s) 0.275167 +time/saving (s) 0.00343923 +time/training (s) 7.12708 +time/epoch (s) 10.1269 +time/total (s) 2618.83 +Epoch -752 +---------------------------------- --------------- +2022-05-10 13:54:26.543384 PDT | [1] Epoch -751 finished +---------------------------------- --------------- +epoch -751 +replay_buffer/size 999033 +trainer/num train calls 250000 +trainer/Policy Loss -19.6446 +trainer/Log Pis Mean 24.7397 +trainer/Log Pis Std 14.2019 +trainer/Log Pis Max 74.2782 +trainer/Log Pis Min -8.98356 +trainer/policy/mean Mean -0.024907 +trainer/policy/mean Std 0.90957 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85532 +trainer/policy/normal/std Std 0.579838 +trainer/policy/normal/std Max 4.89011 +trainer/policy/normal/std Min 0.477727 +trainer/policy/normal/log_std Mean 1.02138 +trainer/policy/normal/log_std Std 0.260091 +trainer/policy/normal/log_std Max 1.58721 +trainer/policy/normal/log_std Min -0.738715 +eval/num steps total 249303 +eval/num paths total 250 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.205134 +eval/Actions Std 0.895278 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72217 +time/logging (s) 0.00368424 +time/sampling batch (s) 0.27444 +time/saving (s) 0.00344885 +time/training (s) 6.96838 +time/epoch (s) 9.97212 +time/total (s) 2628.8 +Epoch -751 +---------------------------------- --------------- +2022-05-10 13:54:36.211657 PDT | [1] Epoch -750 finished +---------------------------------- --------------- +epoch -750 +replay_buffer/size 999033 +trainer/num train calls 251000 +trainer/Policy Loss -19.2366 +trainer/Log Pis Mean 24.9183 +trainer/Log Pis Std 13.8143 +trainer/Log Pis Max 87.6417 +trainer/Log Pis Min -5.61702 +trainer/policy/mean Mean -0.0458258 +trainer/policy/mean Std 0.901913 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88487 +trainer/policy/normal/std Std 0.621282 +trainer/policy/normal/std Max 5.78314 +trainer/policy/normal/std Min 0.359908 +trainer/policy/normal/log_std Mean 1.02585 +trainer/policy/normal/log_std Std 0.292618 +trainer/policy/normal/log_std Max 1.75495 +trainer/policy/normal/log_std Min -1.02191 +eval/num steps total 250303 +eval/num paths total 251 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0703034 +eval/Actions Std 0.907102 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28719 +time/logging (s) 0.00421581 +time/sampling batch (s) 0.526689 +time/saving (s) 0.00387401 +time/training (s) 6.8261 +time/epoch (s) 9.64807 +time/total (s) 2638.45 +Epoch -750 +---------------------------------- --------------- +2022-05-10 13:54:46.199667 PDT | [1] Epoch -749 finished +---------------------------------- --------------- +epoch -749 +replay_buffer/size 999033 +trainer/num train calls 252000 +trainer/Policy Loss -18.5876 +trainer/Log Pis Mean 23.5165 +trainer/Log Pis Std 12.2142 +trainer/Log Pis Max 62.9294 +trainer/Log Pis Min -9.93425 +trainer/policy/mean Mean -0.0101001 +trainer/policy/mean Std 0.899752 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.79357 +trainer/policy/normal/std Std 0.601583 +trainer/policy/normal/std Max 5.44866 +trainer/policy/normal/std Min 0.355686 +trainer/policy/normal/log_std Mean 0.994712 +trainer/policy/normal/log_std Std 0.285451 +trainer/policy/normal/log_std Max 1.69537 +trainer/policy/normal/log_std Min -1.03371 +eval/num steps total 251303 +eval/num paths total 252 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0189251 +eval/Actions Std 0.903592 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51589 +time/logging (s) 0.00387816 +time/sampling batch (s) 0.278371 +time/saving (s) 0.00375642 +time/training (s) 7.16447 +time/epoch (s) 9.96637 +time/total (s) 2648.42 +Epoch -749 +---------------------------------- --------------- +2022-05-10 13:54:56.085560 PDT | [1] Epoch -748 finished +---------------------------------- --------------- +epoch -748 +replay_buffer/size 999033 +trainer/num train calls 253000 +trainer/Policy Loss -19.6258 +trainer/Log Pis Mean 24.12 +trainer/Log Pis Std 13.3329 +trainer/Log Pis Max 76.8886 +trainer/Log Pis Min -11.2875 +trainer/policy/mean Mean -0.0284153 +trainer/policy/mean Std 0.906431 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8553 +trainer/policy/normal/std Std 0.608546 +trainer/policy/normal/std Max 5.35872 +trainer/policy/normal/std Min 0.347973 +trainer/policy/normal/log_std Mean 1.01798 +trainer/policy/normal/log_std Std 0.277035 +trainer/policy/normal/log_std Max 1.67872 +trainer/policy/normal/log_std Min -1.05563 +eval/num steps total 252303 +eval/num paths total 253 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.127937 +eval/Actions Std 0.91247 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59073 +time/logging (s) 0.00380503 +time/sampling batch (s) 0.279092 +time/saving (s) 0.00348245 +time/training (s) 6.98751 +time/epoch (s) 9.86462 +time/total (s) 2658.29 +Epoch -748 +---------------------------------- --------------- +2022-05-10 13:55:06.655075 PDT | [1] Epoch -747 finished +---------------------------------- --------------- +epoch -747 +replay_buffer/size 999033 +trainer/num train calls 254000 +trainer/Policy Loss -19.8926 +trainer/Log Pis Mean 24.6819 +trainer/Log Pis Std 13.486 +trainer/Log Pis Max 66.3813 +trainer/Log Pis Min -6.04467 +trainer/policy/mean Mean -0.0476914 +trainer/policy/mean Std 0.909927 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84066 +trainer/policy/normal/std Std 0.587976 +trainer/policy/normal/std Max 5.65108 +trainer/policy/normal/std Min 0.397787 +trainer/policy/normal/log_std Mean 1.01441 +trainer/policy/normal/log_std Std 0.270872 +trainer/policy/normal/log_std Max 1.73185 +trainer/policy/normal/log_std Min -0.921839 +eval/num steps total 253303 +eval/num paths total 254 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.25643 +eval/Actions Std 0.878648 +eval/Actions Max 0.999991 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74481 +time/logging (s) 0.00378597 +time/sampling batch (s) 0.278955 +time/saving (s) 0.00342254 +time/training (s) 7.51763 +time/epoch (s) 10.5486 +time/total (s) 2668.84 +Epoch -747 +---------------------------------- --------------- +2022-05-10 13:55:17.074799 PDT | [1] Epoch -746 finished +---------------------------------- --------------- +epoch -746 +replay_buffer/size 999033 +trainer/num train calls 255000 +trainer/Policy Loss -19.3288 +trainer/Log Pis Mean 24.3651 +trainer/Log Pis Std 12.9709 +trainer/Log Pis Max 62.7451 +trainer/Log Pis Min -9.34759 +trainer/policy/mean Mean -0.0358101 +trainer/policy/mean Std 0.905604 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.86445 +trainer/policy/normal/std Std 0.602454 +trainer/policy/normal/std Max 5.34935 +trainer/policy/normal/std Min 0.371725 +trainer/policy/normal/log_std Mean 1.02092 +trainer/policy/normal/log_std Std 0.281018 +trainer/policy/normal/log_std Max 1.67698 +trainer/policy/normal/log_std Min -0.9896 +eval/num steps total 254303 +eval/num paths total 255 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.305585 +eval/Actions Std 0.843774 +eval/Actions Max 0.999995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37181 +time/logging (s) 0.003715 +time/sampling batch (s) 0.780247 +time/saving (s) 0.00342244 +time/training (s) 7.23948 +time/epoch (s) 10.3987 +time/total (s) 2679.24 +Epoch -746 +---------------------------------- --------------- +2022-05-10 13:55:26.454432 PDT | [1] Epoch -745 finished +---------------------------------- --------------- +epoch -745 +replay_buffer/size 999033 +trainer/num train calls 256000 +trainer/Policy Loss -18.3955 +trainer/Log Pis Mean 24.0989 +trainer/Log Pis Std 13.3401 +trainer/Log Pis Max 69.587 +trainer/Log Pis Min -11.5103 +trainer/policy/mean Mean -0.0361258 +trainer/policy/mean Std 0.901886 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8211 +trainer/policy/normal/std Std 0.594821 +trainer/policy/normal/std Max 5.52231 +trainer/policy/normal/std Min 0.43574 +trainer/policy/normal/log_std Mean 1.00665 +trainer/policy/normal/log_std Std 0.274197 +trainer/policy/normal/log_std Max 1.7088 +trainer/policy/normal/log_std Min -0.830709 +eval/num steps total 255303 +eval/num paths total 256 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10871 +eval/Actions Std 0.900777 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63234 +time/logging (s) 0.00370759 +time/sampling batch (s) 0.27849 +time/saving (s) 0.00345048 +time/training (s) 6.44069 +time/epoch (s) 9.35868 +time/total (s) 2688.6 +Epoch -745 +---------------------------------- --------------- +2022-05-10 13:55:35.513746 PDT | [1] Epoch -744 finished +---------------------------------- --------------- +epoch -744 +replay_buffer/size 999033 +trainer/num train calls 257000 +trainer/Policy Loss -19.5511 +trainer/Log Pis Mean 25.2123 +trainer/Log Pis Std 13.587 +trainer/Log Pis Max 71.023 +trainer/Log Pis Min -7.03199 +trainer/policy/mean Mean -0.026651 +trainer/policy/mean Std 0.901247 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82827 +trainer/policy/normal/std Std 0.608522 +trainer/policy/normal/std Max 5.49162 +trainer/policy/normal/std Min 0.375915 +trainer/policy/normal/log_std Mean 1.00715 +trainer/policy/normal/log_std Std 0.284561 +trainer/policy/normal/log_std Max 1.70322 +trainer/policy/normal/log_std Min -0.978391 +eval/num steps total 256303 +eval/num paths total 257 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.147093 +eval/Actions Std 0.924311 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7515 +time/logging (s) 0.00413196 +time/sampling batch (s) 0.278569 +time/saving (s) 0.0041661 +time/training (s) 6.00037 +time/epoch (s) 9.03874 +time/total (s) 2697.64 +Epoch -744 +---------------------------------- --------------- +2022-05-10 13:55:46.184934 PDT | [1] Epoch -743 finished +---------------------------------- --------------- +epoch -743 +replay_buffer/size 999033 +trainer/num train calls 258000 +trainer/Policy Loss -19.5655 +trainer/Log Pis Mean 24.4537 +trainer/Log Pis Std 12.6707 +trainer/Log Pis Max 61.9435 +trainer/Log Pis Min -6.50329 +trainer/policy/mean Mean -0.0455693 +trainer/policy/mean Std 0.910776 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82314 +trainer/policy/normal/std Std 0.603852 +trainer/policy/normal/std Max 6.45742 +trainer/policy/normal/std Min 0.281037 +trainer/policy/normal/log_std Mean 1.00588 +trainer/policy/normal/log_std Std 0.28267 +trainer/policy/normal/log_std Max 1.86523 +trainer/policy/normal/log_std Min -1.26927 +eval/num steps total 257303 +eval/num paths total 258 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103183 +eval/Actions Std 0.913406 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45434 +time/logging (s) 0.00442662 +time/sampling batch (s) 0.530043 +time/saving (s) 0.00371176 +time/training (s) 7.65736 +time/epoch (s) 10.6499 +time/total (s) 2708.3 +Epoch -743 +---------------------------------- --------------- +2022-05-10 13:55:55.445330 PDT | [1] Epoch -742 finished +---------------------------------- --------------- +epoch -742 +replay_buffer/size 999033 +trainer/num train calls 259000 +trainer/Policy Loss -19.335 +trainer/Log Pis Mean 24.0774 +trainer/Log Pis Std 12.912 +trainer/Log Pis Max 66.2635 +trainer/Log Pis Min -5.65583 +trainer/policy/mean Mean -0.0410907 +trainer/policy/mean Std 0.902794 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82582 +trainer/policy/normal/std Std 0.590745 +trainer/policy/normal/std Max 5.04241 +trainer/policy/normal/std Min 0.371152 +trainer/policy/normal/log_std Mean 1.00836 +trainer/policy/normal/log_std Std 0.274395 +trainer/policy/normal/log_std Max 1.61788 +trainer/policy/normal/log_std Min -0.991143 +eval/num steps total 258303 +eval/num paths total 259 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113806 +eval/Actions Std 0.912207 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54919 +time/logging (s) 0.00376869 +time/sampling batch (s) 0.282051 +time/saving (s) 0.00356391 +time/training (s) 6.39951 +time/epoch (s) 9.23808 +time/total (s) 2717.54 +Epoch -742 +---------------------------------- --------------- +2022-05-10 13:56:04.826850 PDT | [1] Epoch -741 finished +---------------------------------- --------------- +epoch -741 +replay_buffer/size 999033 +trainer/num train calls 260000 +trainer/Policy Loss -18.9614 +trainer/Log Pis Mean 23.8643 +trainer/Log Pis Std 13.5364 +trainer/Log Pis Max 68.6698 +trainer/Log Pis Min -5.6491 +trainer/policy/mean Mean -0.0203273 +trainer/policy/mean Std 0.905042 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89556 +trainer/policy/normal/std Std 0.627418 +trainer/policy/normal/std Max 6.78363 +trainer/policy/normal/std Min 0.402025 +trainer/policy/normal/log_std Mean 1.03046 +trainer/policy/normal/log_std Std 0.285903 +trainer/policy/normal/log_std Max 1.91451 +trainer/policy/normal/log_std Min -0.911241 +eval/num steps total 259303 +eval/num paths total 260 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106294 +eval/Actions Std 0.910509 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54199 +time/logging (s) 0.00371063 +time/sampling batch (s) 0.277891 +time/saving (s) 0.00340762 +time/training (s) 6.53352 +time/epoch (s) 9.36052 +time/total (s) 2726.9 +Epoch -741 +---------------------------------- --------------- +2022-05-10 13:56:14.854038 PDT | [1] Epoch -740 finished +---------------------------------- --------------- +epoch -740 +replay_buffer/size 999033 +trainer/num train calls 261000 +trainer/Policy Loss -19.834 +trainer/Log Pis Mean 25.4913 +trainer/Log Pis Std 13.9211 +trainer/Log Pis Max 64.647 +trainer/Log Pis Min -11.3371 +trainer/policy/mean Mean -0.0519143 +trainer/policy/mean Std 0.911902 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88639 +trainer/policy/normal/std Std 0.594022 +trainer/policy/normal/std Max 5.53891 +trainer/policy/normal/std Min 0.390969 +trainer/policy/normal/log_std Mean 1.03074 +trainer/policy/normal/log_std Std 0.269 +trainer/policy/normal/log_std Max 1.7118 +trainer/policy/normal/log_std Min -0.939126 +eval/num steps total 260303 +eval/num paths total 261 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.148984 +eval/Actions Std 0.881361 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71275 +time/logging (s) 0.00376631 +time/sampling batch (s) 0.528108 +time/saving (s) 0.00342295 +time/training (s) 6.75814 +time/epoch (s) 10.0062 +time/total (s) 2736.91 +Epoch -740 +---------------------------------- --------------- +2022-05-10 13:56:24.856628 PDT | [1] Epoch -739 finished +---------------------------------- --------------- +epoch -739 +replay_buffer/size 999033 +trainer/num train calls 262000 +trainer/Policy Loss -19.927 +trainer/Log Pis Mean 24.2016 +trainer/Log Pis Std 13.631 +trainer/Log Pis Max 63.939 +trainer/Log Pis Min -8.75361 +trainer/policy/mean Mean -0.0521693 +trainer/policy/mean Std 0.902336 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81231 +trainer/policy/normal/std Std 0.618783 +trainer/policy/normal/std Max 5.18451 +trainer/policy/normal/std Min 0.255467 +trainer/policy/normal/log_std Mean 0.999366 +trainer/policy/normal/log_std Std 0.295801 +trainer/policy/normal/log_std Max 1.64567 +trainer/policy/normal/log_std Min -1.36466 +eval/num steps total 261303 +eval/num paths total 262 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113818 +eval/Actions Std 0.906357 +eval/Actions Max 0.999984 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46216 +time/logging (s) 0.00542918 +time/sampling batch (s) 0.529176 +time/saving (s) 0.00578931 +time/training (s) 6.98069 +time/epoch (s) 9.98325 +time/total (s) 2746.9 +Epoch -739 +---------------------------------- --------------- +2022-05-10 13:56:34.837920 PDT | [1] Epoch -738 finished +---------------------------------- --------------- +epoch -738 +replay_buffer/size 999033 +trainer/num train calls 263000 +trainer/Policy Loss -19.5924 +trainer/Log Pis Mean 24.8539 +trainer/Log Pis Std 13.3756 +trainer/Log Pis Max 79.1973 +trainer/Log Pis Min -8.21372 +trainer/policy/mean Mean -0.0317244 +trainer/policy/mean Std 0.90303 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81575 +trainer/policy/normal/std Std 0.560901 +trainer/policy/normal/std Max 5.25729 +trainer/policy/normal/std Min 0.397131 +trainer/policy/normal/log_std Mean 1.00748 +trainer/policy/normal/log_std Std 0.262659 +trainer/policy/normal/log_std Max 1.65962 +trainer/policy/normal/log_std Min -0.923489 +eval/num steps total 262303 +eval/num paths total 263 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120432 +eval/Actions Std 0.899971 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52973 +time/logging (s) 0.00386639 +time/sampling batch (s) 0.530428 +time/saving (s) 0.00372202 +time/training (s) 6.88913 +time/epoch (s) 9.95687 +time/total (s) 2756.86 +Epoch -738 +---------------------------------- --------------- +2022-05-10 13:56:45.583224 PDT | [1] Epoch -737 finished +---------------------------------- --------------- +epoch -737 +replay_buffer/size 999033 +trainer/num train calls 264000 +trainer/Policy Loss -18.1073 +trainer/Log Pis Mean 23.5973 +trainer/Log Pis Std 12.7105 +trainer/Log Pis Max 64.5127 +trainer/Log Pis Min -11.3543 +trainer/policy/mean Mean -0.0369473 +trainer/policy/mean Std 0.907718 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83709 +trainer/policy/normal/std Std 0.573379 +trainer/policy/normal/std Max 5.53645 +trainer/policy/normal/std Min 0.364622 +trainer/policy/normal/log_std Mean 1.01584 +trainer/policy/normal/log_std Std 0.254096 +trainer/policy/normal/log_std Max 1.71135 +trainer/policy/normal/log_std Min -1.00889 +eval/num steps total 263303 +eval/num paths total 264 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00443858 +eval/Actions Std 0.901636 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6778 +time/logging (s) 0.00373097 +time/sampling batch (s) 0.276098 +time/saving (s) 0.00342263 +time/training (s) 7.76319 +time/epoch (s) 10.7242 +time/total (s) 2767.59 +Epoch -737 +---------------------------------- --------------- +2022-05-10 13:56:55.667141 PDT | [1] Epoch -736 finished +---------------------------------- --------------- +epoch -736 +replay_buffer/size 999033 +trainer/num train calls 265000 +trainer/Policy Loss -21.2873 +trainer/Log Pis Mean 24.7691 +trainer/Log Pis Std 13.4078 +trainer/Log Pis Max 65.0554 +trainer/Log Pis Min -9.21092 +trainer/policy/mean Mean -0.0376366 +trainer/policy/mean Std 0.909189 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86607 +trainer/policy/normal/std Std 0.625239 +trainer/policy/normal/std Max 5.44417 +trainer/policy/normal/std Min 0.291959 +trainer/policy/normal/log_std Mean 1.01768 +trainer/policy/normal/log_std Std 0.302456 +trainer/policy/normal/log_std Max 1.69454 +trainer/policy/normal/log_std Min -1.23114 +eval/num steps total 264303 +eval/num paths total 265 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0135788 +eval/Actions Std 0.898798 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59142 +time/logging (s) 0.00423768 +time/sampling batch (s) 0.775173 +time/saving (s) 0.00394583 +time/training (s) 6.68876 +time/epoch (s) 10.0635 +time/total (s) 2777.65 +Epoch -736 +---------------------------------- --------------- +2022-05-10 13:57:06.344098 PDT | [1] Epoch -735 finished +---------------------------------- --------------- +epoch -735 +replay_buffer/size 999033 +trainer/num train calls 266000 +trainer/Policy Loss -19.1344 +trainer/Log Pis Mean 24.5247 +trainer/Log Pis Std 13.8187 +trainer/Log Pis Max 70.4076 +trainer/Log Pis Min -11.9668 +trainer/policy/mean Mean -0.0358981 +trainer/policy/mean Std 0.905834 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.82942 +trainer/policy/normal/std Std 0.597965 +trainer/policy/normal/std Max 5.75416 +trainer/policy/normal/std Min 0.42064 +trainer/policy/normal/log_std Mean 1.00905 +trainer/policy/normal/log_std Std 0.277513 +trainer/policy/normal/log_std Max 1.74992 +trainer/policy/normal/log_std Min -0.865978 +eval/num steps total 265303 +eval/num paths total 266 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.200836 +eval/Actions Std 0.884326 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50407 +time/logging (s) 0.00382061 +time/sampling batch (s) 0.78592 +time/saving (s) 0.00359946 +time/training (s) 7.35725 +time/epoch (s) 10.6547 +time/total (s) 2788.31 +Epoch -735 +---------------------------------- --------------- +2022-05-10 13:57:16.346285 PDT | [1] Epoch -734 finished +---------------------------------- --------------- +epoch -734 +replay_buffer/size 999033 +trainer/num train calls 267000 +trainer/Policy Loss -19.3483 +trainer/Log Pis Mean 24.8579 +trainer/Log Pis Std 13.0847 +trainer/Log Pis Max 61.4032 +trainer/Log Pis Min -9.71532 +trainer/policy/mean Mean -0.038441 +trainer/policy/mean Std 0.909621 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83987 +trainer/policy/normal/std Std 0.60267 +trainer/policy/normal/std Max 5.29725 +trainer/policy/normal/std Min 0.372974 +trainer/policy/normal/log_std Mean 1.01157 +trainer/policy/normal/log_std Std 0.284845 +trainer/policy/normal/log_std Max 1.66719 +trainer/policy/normal/log_std Min -0.986247 +eval/num steps total 266303 +eval/num paths total 267 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.185668 +eval/Actions Std 0.914263 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52153 +time/logging (s) 0.00384862 +time/sampling batch (s) 0.282468 +time/saving (s) 0.00347173 +time/training (s) 7.16955 +time/epoch (s) 9.98087 +time/total (s) 2798.29 +Epoch -734 +---------------------------------- --------------- +2022-05-10 13:57:26.914162 PDT | [1] Epoch -733 finished +---------------------------------- --------------- +epoch -733 +replay_buffer/size 999033 +trainer/num train calls 268000 +trainer/Policy Loss -19.6843 +trainer/Log Pis Mean 24.2569 +trainer/Log Pis Std 13.1695 +trainer/Log Pis Max 65.8895 +trainer/Log Pis Min -6.47872 +trainer/policy/mean Mean -0.0231017 +trainer/policy/mean Std 0.904529 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7727 +trainer/policy/normal/std Std 0.597563 +trainer/policy/normal/std Max 5.60397 +trainer/policy/normal/std Min 0.401204 +trainer/policy/normal/log_std Mean 0.986765 +trainer/policy/normal/log_std Std 0.2876 +trainer/policy/normal/log_std Max 1.72347 +trainer/policy/normal/log_std Min -0.913284 +eval/num steps total 267303 +eval/num paths total 268 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.179945 +eval/Actions Std 0.882601 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60727 +time/logging (s) 0.00424249 +time/sampling batch (s) 0.532107 +time/saving (s) 0.00399254 +time/training (s) 7.39922 +time/epoch (s) 10.5468 +time/total (s) 2808.84 +Epoch -733 +---------------------------------- --------------- +2022-05-10 13:57:36.317923 PDT | [1] Epoch -732 finished +---------------------------------- --------------- +epoch -732 +replay_buffer/size 999033 +trainer/num train calls 269000 +trainer/Policy Loss -19.9542 +trainer/Log Pis Mean 24.7172 +trainer/Log Pis Std 13.1206 +trainer/Log Pis Max 63.4046 +trainer/Log Pis Min -6.62672 +trainer/policy/mean Mean -0.0357306 +trainer/policy/mean Std 0.903357 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.83486 +trainer/policy/normal/std Std 0.619153 +trainer/policy/normal/std Max 5.39295 +trainer/policy/normal/std Min 0.34416 +trainer/policy/normal/log_std Mean 1.00763 +trainer/policy/normal/log_std Std 0.295194 +trainer/policy/normal/log_std Max 1.68509 +trainer/policy/normal/log_std Min -1.06665 +eval/num steps total 268303 +eval/num paths total 269 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0317883 +eval/Actions Std 0.904873 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5663 +time/logging (s) 0.00382486 +time/sampling batch (s) 0.278419 +time/saving (s) 0.00348335 +time/training (s) 6.52998 +time/epoch (s) 9.382 +time/total (s) 2818.23 +Epoch -732 +---------------------------------- --------------- +2022-05-10 13:57:46.601024 PDT | [1] Epoch -731 finished +---------------------------------- --------------- +epoch -731 +replay_buffer/size 999033 +trainer/num train calls 270000 +trainer/Policy Loss -19.1819 +trainer/Log Pis Mean 23.3714 +trainer/Log Pis Std 12.4608 +trainer/Log Pis Max 58.7489 +trainer/Log Pis Min -4.28129 +trainer/policy/mean Mean -0.0339146 +trainer/policy/mean Std 0.903934 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85552 +trainer/policy/normal/std Std 0.608017 +trainer/policy/normal/std Max 5.67176 +trainer/policy/normal/std Min 0.374159 +trainer/policy/normal/log_std Mean 1.01696 +trainer/policy/normal/log_std Std 0.285009 +trainer/policy/normal/log_std Max 1.7355 +trainer/policy/normal/log_std Min -0.983076 +eval/num steps total 269303 +eval/num paths total 270 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.333705 +eval/Actions Std 0.869772 +eval/Actions Max 0.999996 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76595 +time/logging (s) 0.00390228 +time/sampling batch (s) 0.280158 +time/saving (s) 0.00342348 +time/training (s) 7.20851 +time/epoch (s) 10.2619 +time/total (s) 2828.49 +Epoch -731 +---------------------------------- --------------- +2022-05-10 13:57:57.013038 PDT | [1] Epoch -730 finished +---------------------------------- --------------- +epoch -730 +replay_buffer/size 999033 +trainer/num train calls 271000 +trainer/Policy Loss -19.3039 +trainer/Log Pis Mean 24.633 +trainer/Log Pis Std 13.7565 +trainer/Log Pis Max 76.425 +trainer/Log Pis Min -9.06383 +trainer/policy/mean Mean -0.0601332 +trainer/policy/mean Std 0.90591 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8055 +trainer/policy/normal/std Std 0.599238 +trainer/policy/normal/std Max 5.94236 +trainer/policy/normal/std Min 0.352249 +trainer/policy/normal/log_std Mean 0.999961 +trainer/policy/normal/log_std Std 0.280099 +trainer/policy/normal/log_std Max 1.78211 +trainer/policy/normal/log_std Min -1.04342 +eval/num steps total 270303 +eval/num paths total 271 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0988772 +eval/Actions Std 0.908844 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73389 +time/logging (s) 0.00414829 +time/sampling batch (s) 0.528835 +time/saving (s) 0.00383321 +time/training (s) 7.12044 +time/epoch (s) 10.3911 +time/total (s) 2838.89 +Epoch -730 +---------------------------------- --------------- +2022-05-10 13:58:06.814013 PDT | [1] Epoch -729 finished +---------------------------------- --------------- +epoch -729 +replay_buffer/size 999033 +trainer/num train calls 272000 +trainer/Policy Loss -19.2727 +trainer/Log Pis Mean 24.4449 +trainer/Log Pis Std 12.6546 +trainer/Log Pis Max 59.6919 +trainer/Log Pis Min -11.4363 +trainer/policy/mean Mean -0.0313171 +trainer/policy/mean Std 0.903625 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83064 +trainer/policy/normal/std Std 0.63198 +trainer/policy/normal/std Max 6.76929 +trainer/policy/normal/std Min 0.346225 +trainer/policy/normal/log_std Mean 1.00557 +trainer/policy/normal/log_std Std 0.296291 +trainer/policy/normal/log_std Max 1.9124 +trainer/policy/normal/log_std Min -1.06067 +eval/num steps total 271303 +eval/num paths total 272 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.259073 +eval/Actions Std 0.876559 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62073 +time/logging (s) 0.00391957 +time/sampling batch (s) 0.529395 +time/saving (s) 0.00379415 +time/training (s) 6.62134 +time/epoch (s) 9.77918 +time/total (s) 2848.67 +Epoch -729 +---------------------------------- --------------- +2022-05-10 13:58:17.433165 PDT | [1] Epoch -728 finished +---------------------------------- --------------- +epoch -728 +replay_buffer/size 999033 +trainer/num train calls 273000 +trainer/Policy Loss -18.5419 +trainer/Log Pis Mean 24.4216 +trainer/Log Pis Std 14.0235 +trainer/Log Pis Max 93.0726 +trainer/Log Pis Min -9.58589 +trainer/policy/mean Mean -0.0371878 +trainer/policy/mean Std 0.905945 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82711 +trainer/policy/normal/std Std 0.634509 +trainer/policy/normal/std Max 7.0902 +trainer/policy/normal/std Min 0.268093 +trainer/policy/normal/log_std Mean 1.00392 +trainer/policy/normal/log_std Std 0.29776 +trainer/policy/normal/log_std Max 1.95871 +trainer/policy/normal/log_std Min -1.31642 +eval/num steps total 272303 +eval/num paths total 273 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132825 +eval/Actions Std 0.831194 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54455 +time/logging (s) 0.00368294 +time/sampling batch (s) 0.281274 +time/saving (s) 0.00346383 +time/training (s) 7.76442 +time/epoch (s) 10.5974 +time/total (s) 2859.27 +Epoch -728 +---------------------------------- --------------- +2022-05-10 13:58:27.351836 PDT | [1] Epoch -727 finished +---------------------------------- --------------- +epoch -727 +replay_buffer/size 999033 +trainer/num train calls 274000 +trainer/Policy Loss -19.3224 +trainer/Log Pis Mean 25.0389 +trainer/Log Pis Std 12.6523 +trainer/Log Pis Max 70.1549 +trainer/Log Pis Min -5.25279 +trainer/policy/mean Mean -0.0495697 +trainer/policy/mean Std 0.904941 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85541 +trainer/policy/normal/std Std 0.629008 +trainer/policy/normal/std Max 5.36824 +trainer/policy/normal/std Min 0.365442 +trainer/policy/normal/log_std Mean 1.01484 +trainer/policy/normal/log_std Std 0.293843 +trainer/policy/normal/log_std Max 1.6805 +trainer/policy/normal/log_std Min -1.00665 +eval/num steps total 273303 +eval/num paths total 274 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.403945 +eval/Actions Std 0.858108 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61615 +time/logging (s) 0.00375611 +time/sampling batch (s) 0.282953 +time/saving (s) 0.0034969 +time/training (s) 6.99106 +time/epoch (s) 9.89742 +time/total (s) 2869.17 +Epoch -727 +---------------------------------- --------------- +2022-05-10 13:58:36.890733 PDT | [1] Epoch -726 finished +---------------------------------- --------------- +epoch -726 +replay_buffer/size 999033 +trainer/num train calls 275000 +trainer/Policy Loss -19.1053 +trainer/Log Pis Mean 23.8029 +trainer/Log Pis Std 13.1388 +trainer/Log Pis Max 80.4624 +trainer/Log Pis Min -5.73814 +trainer/policy/mean Mean -0.0208813 +trainer/policy/mean Std 0.9071 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8574 +trainer/policy/normal/std Std 0.619334 +trainer/policy/normal/std Max 5.66416 +trainer/policy/normal/std Min 0.35346 +trainer/policy/normal/log_std Mean 1.01584 +trainer/policy/normal/log_std Std 0.294085 +trainer/policy/normal/log_std Max 1.73416 +trainer/policy/normal/log_std Min -1.03999 +eval/num steps total 274303 +eval/num paths total 275 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00957379 +eval/Actions Std 0.904713 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69471 +time/logging (s) 0.00413054 +time/sampling batch (s) 0.281847 +time/saving (s) 0.00390453 +time/training (s) 6.53326 +time/epoch (s) 9.51786 +time/total (s) 2878.69 +Epoch -726 +---------------------------------- --------------- +2022-05-10 13:58:47.220569 PDT | [1] Epoch -725 finished +---------------------------------- --------------- +epoch -725 +replay_buffer/size 999033 +trainer/num train calls 276000 +trainer/Policy Loss -19.3834 +trainer/Log Pis Mean 26.0034 +trainer/Log Pis Std 13.1771 +trainer/Log Pis Max 70.8624 +trainer/Log Pis Min -9.04635 +trainer/policy/mean Mean -0.0458708 +trainer/policy/mean Std 0.909624 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86725 +trainer/policy/normal/std Std 0.624019 +trainer/policy/normal/std Max 6.241 +trainer/policy/normal/std Min 0.364621 +trainer/policy/normal/log_std Mean 1.02141 +trainer/policy/normal/log_std Std 0.280145 +trainer/policy/normal/log_std Max 1.83114 +trainer/policy/normal/log_std Min -1.0089 +eval/num steps total 275303 +eval/num paths total 276 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0264693 +eval/Actions Std 0.907514 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71827 +time/logging (s) 0.00386794 +time/sampling batch (s) 0.280851 +time/saving (s) 0.00371003 +time/training (s) 7.30123 +time/epoch (s) 10.3079 +time/total (s) 2889 +Epoch -725 +---------------------------------- --------------- +2022-05-10 13:58:57.407514 PDT | [1] Epoch -724 finished +---------------------------------- --------------- +epoch -724 +replay_buffer/size 999033 +trainer/num train calls 277000 +trainer/Policy Loss -20.1931 +trainer/Log Pis Mean 24.457 +trainer/Log Pis Std 12.8987 +trainer/Log Pis Max 82.9589 +trainer/Log Pis Min -8.55089 +trainer/policy/mean Mean -0.0145858 +trainer/policy/mean Std 0.906129 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82769 +trainer/policy/normal/std Std 0.608313 +trainer/policy/normal/std Max 5.40997 +trainer/policy/normal/std Min 0.325994 +trainer/policy/normal/log_std Mean 1.00701 +trainer/policy/normal/log_std Std 0.284526 +trainer/policy/normal/log_std Max 1.68824 +trainer/policy/normal/log_std Min -1.12088 +eval/num steps total 276303 +eval/num paths total 277 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0181304 +eval/Actions Std 0.929118 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79387 +time/logging (s) 0.00394572 +time/sampling batch (s) 0.279369 +time/saving (s) 0.00361014 +time/training (s) 7.08502 +time/epoch (s) 10.1658 +time/total (s) 2899.17 +Epoch -724 +---------------------------------- --------------- +2022-05-10 13:59:07.977513 PDT | [1] Epoch -723 finished +---------------------------------- --------------- +epoch -723 +replay_buffer/size 999033 +trainer/num train calls 278000 +trainer/Policy Loss -19.1581 +trainer/Log Pis Mean 24.684 +trainer/Log Pis Std 13.0857 +trainer/Log Pis Max 66.4962 +trainer/Log Pis Min -6.14067 +trainer/policy/mean Mean -0.0245076 +trainer/policy/mean Std 0.904596 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78093 +trainer/policy/normal/std Std 0.619237 +trainer/policy/normal/std Max 5.79044 +trainer/policy/normal/std Min 0.424206 +trainer/policy/normal/log_std Mean 0.987485 +trainer/policy/normal/log_std Std 0.297668 +trainer/policy/normal/log_std Max 1.75621 +trainer/policy/normal/log_std Min -0.857536 +eval/num steps total 277303 +eval/num paths total 278 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.271413 +eval/Actions Std 0.881864 +eval/Actions Max 0.999975 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5644 +time/logging (s) 0.00417832 +time/sampling batch (s) 0.531148 +time/saving (s) 0.003988 +time/training (s) 7.44506 +time/epoch (s) 10.5488 +time/total (s) 2909.72 +Epoch -723 +---------------------------------- --------------- +2022-05-10 13:59:19.518388 PDT | [1] Epoch -722 finished +---------------------------------- --------------- +epoch -722 +replay_buffer/size 999033 +trainer/num train calls 279000 +trainer/Policy Loss -20.3249 +trainer/Log Pis Mean 24.9581 +trainer/Log Pis Std 13.4319 +trainer/Log Pis Max 66.4369 +trainer/Log Pis Min -6.78536 +trainer/policy/mean Mean -0.062645 +trainer/policy/mean Std 0.905489 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84987 +trainer/policy/normal/std Std 0.603183 +trainer/policy/normal/std Max 5.80189 +trainer/policy/normal/std Min 0.393406 +trainer/policy/normal/log_std Mean 1.01494 +trainer/policy/normal/log_std Std 0.286037 +trainer/policy/normal/log_std Max 1.75818 +trainer/policy/normal/log_std Min -0.932914 +eval/num steps total 278303 +eval/num paths total 279 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00852882 +eval/Actions Std 0.909026 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63601 +time/logging (s) 0.00379477 +time/sampling batch (s) 0.529466 +time/saving (s) 0.00362638 +time/training (s) 8.34592 +time/epoch (s) 11.5188 +time/total (s) 2921.25 +Epoch -722 +---------------------------------- --------------- +2022-05-10 13:59:29.636354 PDT | [1] Epoch -721 finished +---------------------------------- --------------- +epoch -721 +replay_buffer/size 999033 +trainer/num train calls 280000 +trainer/Policy Loss -20.7428 +trainer/Log Pis Mean 25.405 +trainer/Log Pis Std 13.6557 +trainer/Log Pis Max 77.8961 +trainer/Log Pis Min -10.943 +trainer/policy/mean Mean -0.0453443 +trainer/policy/mean Std 0.910236 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84745 +trainer/policy/normal/std Std 0.615896 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.282176 +trainer/policy/normal/log_std Mean 1.01229 +trainer/policy/normal/log_std Std 0.296027 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.26522 +eval/num steps total 279303 +eval/num paths total 280 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.509914 +eval/Actions Std 0.681384 +eval/Actions Max 0.999996 +eval/Actions Min -0.999971 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47602 +time/logging (s) 0.00369124 +time/sampling batch (s) 0.527882 +time/saving (s) 0.00342914 +time/training (s) 7.08578 +time/epoch (s) 10.0968 +time/total (s) 2931.35 +Epoch -721 +---------------------------------- --------------- +2022-05-10 13:59:39.953075 PDT | [1] Epoch -720 finished +---------------------------------- --------------- +epoch -720 +replay_buffer/size 999033 +trainer/num train calls 281000 +trainer/Policy Loss -20.8291 +trainer/Log Pis Mean 24.9243 +trainer/Log Pis Std 13.5339 +trainer/Log Pis Max 75.0096 +trainer/Log Pis Min -8.5061 +trainer/policy/mean Mean -0.0418172 +trainer/policy/mean Std 0.907496 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86706 +trainer/policy/normal/std Std 0.609996 +trainer/policy/normal/std Max 5.7617 +trainer/policy/normal/std Min 0.31214 +trainer/policy/normal/log_std Mean 1.0212 +trainer/policy/normal/log_std Std 0.283763 +trainer/policy/normal/log_std Max 1.75123 +trainer/policy/normal/log_std Min -1.1643 +eval/num steps total 280303 +eval/num paths total 281 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.142727 +eval/Actions Std 0.911749 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60456 +time/logging (s) 0.00393827 +time/sampling batch (s) 0.277632 +time/saving (s) 0.00354044 +time/training (s) 7.40624 +time/epoch (s) 10.2959 +time/total (s) 2941.65 +Epoch -720 +---------------------------------- --------------- +2022-05-10 13:59:50.404916 PDT | [1] Epoch -719 finished +---------------------------------- --------------- +epoch -719 +replay_buffer/size 999033 +trainer/num train calls 282000 +trainer/Policy Loss -19.1775 +trainer/Log Pis Mean 26.4523 +trainer/Log Pis Std 13.4571 +trainer/Log Pis Max 74.2129 +trainer/Log Pis Min -6.67683 +trainer/policy/mean Mean -0.0550402 +trainer/policy/mean Std 0.905368 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.8789 +trainer/policy/normal/std Std 0.613245 +trainer/policy/normal/std Max 5.51511 +trainer/policy/normal/std Min 0.26751 +trainer/policy/normal/log_std Mean 1.02447 +trainer/policy/normal/log_std Std 0.290592 +trainer/policy/normal/log_std Max 1.70749 +trainer/policy/normal/log_std Min -1.3186 +eval/num steps total 281303 +eval/num paths total 282 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.430815 +eval/Actions Std 0.862013 +eval/Actions Max 0.999989 +eval/Actions Min -0.999925 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56442 +time/logging (s) 0.00429456 +time/sampling batch (s) 0.282305 +time/saving (s) 0.00404461 +time/training (s) 7.5756 +time/epoch (s) 10.4307 +time/total (s) 2952.08 +Epoch -719 +---------------------------------- --------------- +2022-05-10 14:00:00.041481 PDT | [1] Epoch -718 finished +---------------------------------- --------------- +epoch -718 +replay_buffer/size 999033 +trainer/num train calls 283000 +trainer/Policy Loss -20.2659 +trainer/Log Pis Mean 24.7251 +trainer/Log Pis Std 13.8401 +trainer/Log Pis Max 61.4438 +trainer/Log Pis Min -8.40979 +trainer/policy/mean Mean -0.0418012 +trainer/policy/mean Std 0.907271 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8219 +trainer/policy/normal/std Std 0.60141 +trainer/policy/normal/std Max 6.28417 +trainer/policy/normal/std Min 0.36328 +trainer/policy/normal/log_std Mean 1.00601 +trainer/policy/normal/log_std Std 0.279188 +trainer/policy/normal/log_std Max 1.83803 +trainer/policy/normal/log_std Min -1.01258 +eval/num steps total 282303 +eval/num paths total 283 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0448101 +eval/Actions Std 0.930615 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59626 +time/logging (s) 0.00388223 +time/sampling batch (s) 0.2816 +time/saving (s) 0.00360265 +time/training (s) 6.72885 +time/epoch (s) 9.61419 +time/total (s) 2961.7 +Epoch -718 +---------------------------------- --------------- +2022-05-10 14:00:09.143158 PDT | [1] Epoch -717 finished +---------------------------------- --------------- +epoch -717 +replay_buffer/size 999033 +trainer/num train calls 284000 +trainer/Policy Loss -19.3679 +trainer/Log Pis Mean 23.3643 +trainer/Log Pis Std 13.2827 +trainer/Log Pis Max 60.9919 +trainer/Log Pis Min -10.1795 +trainer/policy/mean Mean -0.0282034 +trainer/policy/mean Std 0.900856 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77662 +trainer/policy/normal/std Std 0.62918 +trainer/policy/normal/std Max 6.78217 +trainer/policy/normal/std Min 0.34961 +trainer/policy/normal/log_std Mean 0.984309 +trainer/policy/normal/log_std Std 0.306436 +trainer/policy/normal/log_std Max 1.9143 +trainer/policy/normal/log_std Min -1.05094 +eval/num steps total 283303 +eval/num paths total 284 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0822005 +eval/Actions Std 0.916268 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64834 +time/logging (s) 0.00379102 +time/sampling batch (s) 0.281141 +time/saving (s) 0.00349368 +time/training (s) 6.14296 +time/epoch (s) 9.07973 +time/total (s) 2970.78 +Epoch -717 +---------------------------------- --------------- +2022-05-10 14:00:19.858470 PDT | [1] Epoch -716 finished +---------------------------------- --------------- +epoch -716 +replay_buffer/size 999033 +trainer/num train calls 285000 +trainer/Policy Loss -18.877 +trainer/Log Pis Mean 25.6014 +trainer/Log Pis Std 13.3105 +trainer/Log Pis Max 82.8134 +trainer/Log Pis Min -7.34741 +trainer/policy/mean Mean -0.0374721 +trainer/policy/mean Std 0.906926 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8313 +trainer/policy/normal/std Std 0.607734 +trainer/policy/normal/std Max 5.41659 +trainer/policy/normal/std Min 0.34592 +trainer/policy/normal/log_std Mean 1.00866 +trainer/policy/normal/log_std Std 0.282701 +trainer/policy/normal/log_std Max 1.68947 +trainer/policy/normal/log_std Min -1.06155 +eval/num steps total 284303 +eval/num paths total 285 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0325328 +eval/Actions Std 0.907966 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73063 +time/logging (s) 0.00387086 +time/sampling batch (s) 0.280292 +time/saving (s) 0.00340368 +time/training (s) 7.67531 +time/epoch (s) 10.6935 +time/total (s) 2981.48 +Epoch -716 +---------------------------------- --------------- +2022-05-10 14:00:29.485772 PDT | [1] Epoch -715 finished +---------------------------------- --------------- +epoch -715 +replay_buffer/size 999033 +trainer/num train calls 286000 +trainer/Policy Loss -19.1707 +trainer/Log Pis Mean 24.6196 +trainer/Log Pis Std 13.1176 +trainer/Log Pis Max 83.0921 +trainer/Log Pis Min -5.66954 +trainer/policy/mean Mean -0.0404254 +trainer/policy/mean Std 0.907424 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78529 +trainer/policy/normal/std Std 0.601488 +trainer/policy/normal/std Max 4.85361 +trainer/policy/normal/std Min 0.37294 +trainer/policy/normal/log_std Mean 0.990856 +trainer/policy/normal/log_std Std 0.290613 +trainer/policy/normal/log_std Max 1.57972 +trainer/policy/normal/log_std Min -0.986337 +eval/num steps total 285303 +eval/num paths total 286 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0728235 +eval/Actions Std 0.900033 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5503 +time/logging (s) 0.00428516 +time/sampling batch (s) 0.281434 +time/saving (s) 0.00384711 +time/training (s) 6.7665 +time/epoch (s) 9.60636 +time/total (s) 2991.09 +Epoch -715 +---------------------------------- --------------- +2022-05-10 14:00:39.520427 PDT | [1] Epoch -714 finished +---------------------------------- --------------- +epoch -714 +replay_buffer/size 999033 +trainer/num train calls 287000 +trainer/Policy Loss -20.0257 +trainer/Log Pis Mean 23.5364 +trainer/Log Pis Std 12.8954 +trainer/Log Pis Max 72.9011 +trainer/Log Pis Min -7.86643 +trainer/policy/mean Mean -0.0551944 +trainer/policy/mean Std 0.902185 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86581 +trainer/policy/normal/std Std 0.631509 +trainer/policy/normal/std Max 5.62441 +trainer/policy/normal/std Min 0.323549 +trainer/policy/normal/log_std Mean 1.01694 +trainer/policy/normal/log_std Std 0.304812 +trainer/policy/normal/log_std Max 1.72712 +trainer/policy/normal/log_std Min -1.12841 +eval/num steps total 286303 +eval/num paths total 287 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.140495 +eval/Actions Std 0.891324 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46787 +time/logging (s) 0.00385072 +time/sampling batch (s) 0.280104 +time/saving (s) 0.00370498 +time/training (s) 7.25691 +time/epoch (s) 10.0124 +time/total (s) 3001.1 +Epoch -714 +---------------------------------- --------------- +2022-05-10 14:00:49.811479 PDT | [1] Epoch -713 finished +---------------------------------- --------------- +epoch -713 +replay_buffer/size 999033 +trainer/num train calls 288000 +trainer/Policy Loss -19.2108 +trainer/Log Pis Mean 24.7576 +trainer/Log Pis Std 12.8504 +trainer/Log Pis Max 62.5249 +trainer/Log Pis Min -6.09541 +trainer/policy/mean Mean -0.0496068 +trainer/policy/mean Std 0.907571 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82442 +trainer/policy/normal/std Std 0.585844 +trainer/policy/normal/std Max 5.32842 +trainer/policy/normal/std Min 0.374252 +trainer/policy/normal/log_std Mean 1.0083 +trainer/policy/normal/log_std Std 0.273225 +trainer/policy/normal/log_std Max 1.67306 +trainer/policy/normal/log_std Min -0.982826 +eval/num steps total 287303 +eval/num paths total 288 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.013449 +eval/Actions Std 0.963539 +eval/Actions Max 0.999981 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69509 +time/logging (s) 0.00374306 +time/sampling batch (s) 0.530193 +time/saving (s) 0.00348965 +time/training (s) 7.03701 +time/epoch (s) 10.2695 +time/total (s) 3011.37 +Epoch -713 +---------------------------------- --------------- +2022-05-10 14:00:58.971187 PDT | [1] Epoch -712 finished +---------------------------------- --------------- +epoch -712 +replay_buffer/size 999033 +trainer/num train calls 289000 +trainer/Policy Loss -20.2063 +trainer/Log Pis Mean 24.8613 +trainer/Log Pis Std 13.6673 +trainer/Log Pis Max 68.4034 +trainer/Log Pis Min -10.1267 +trainer/policy/mean Mean -0.0391431 +trainer/policy/mean Std 0.908526 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87907 +trainer/policy/normal/std Std 0.622116 +trainer/policy/normal/std Max 6.66613 +trainer/policy/normal/std Min 0.299286 +trainer/policy/normal/log_std Mean 1.02474 +trainer/policy/normal/log_std Std 0.286862 +trainer/policy/normal/log_std Max 1.89704 +trainer/policy/normal/log_std Min -1.20635 +eval/num steps total 288303 +eval/num paths total 289 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.344339 +eval/Actions Std 0.907564 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57431 +time/logging (s) 0.00384781 +time/sampling batch (s) 0.279476 +time/saving (s) 0.00344204 +time/training (s) 6.27767 +time/epoch (s) 9.13875 +time/total (s) 3020.52 +Epoch -712 +---------------------------------- --------------- +2022-05-10 14:01:07.770043 PDT | [1] Epoch -711 finished +---------------------------------- --------------- +epoch -711 +replay_buffer/size 999033 +trainer/num train calls 290000 +trainer/Policy Loss -20.4582 +trainer/Log Pis Mean 24.4145 +trainer/Log Pis Std 13.0528 +trainer/Log Pis Max 65.2107 +trainer/Log Pis Min -8.09496 +trainer/policy/mean Mean -0.0400235 +trainer/policy/mean Std 0.905209 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87376 +trainer/policy/normal/std Std 0.633584 +trainer/policy/normal/std Max 6.03537 +trainer/policy/normal/std Min 0.331074 +trainer/policy/normal/log_std Mean 1.02126 +trainer/policy/normal/log_std Std 0.294345 +trainer/policy/normal/log_std Max 1.79764 +trainer/policy/normal/log_std Min -1.10541 +eval/num steps total 289303 +eval/num paths total 290 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.557509 +eval/Actions Std 0.670166 +eval/Actions Max 0.999982 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49186 +time/logging (s) 0.00385134 +time/sampling batch (s) 0.278989 +time/saving (s) 0.00352577 +time/training (s) 5.99967 +time/epoch (s) 8.7779 +time/total (s) 3029.3 +Epoch -711 +---------------------------------- --------------- +2022-05-10 14:01:18.225368 PDT | [1] Epoch -710 finished +---------------------------------- --------------- +epoch -710 +replay_buffer/size 999033 +trainer/num train calls 291000 +trainer/Policy Loss -20.0422 +trainer/Log Pis Mean 24.4779 +trainer/Log Pis Std 12.5163 +trainer/Log Pis Max 66.6799 +trainer/Log Pis Min -6.13755 +trainer/policy/mean Mean -0.0380443 +trainer/policy/mean Std 0.902756 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83958 +trainer/policy/normal/std Std 0.621585 +trainer/policy/normal/std Max 5.50017 +trainer/policy/normal/std Min 0.316588 +trainer/policy/normal/log_std Mean 1.0085 +trainer/policy/normal/log_std Std 0.299804 +trainer/policy/normal/log_std Max 1.70478 +trainer/policy/normal/log_std Min -1.15015 +eval/num steps total 290303 +eval/num paths total 291 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.302533 +eval/Actions Std 0.836764 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57959 +time/logging (s) 0.00425145 +time/sampling batch (s) 0.281817 +time/saving (s) 0.00405112 +time/training (s) 7.56458 +time/epoch (s) 10.4343 +time/total (s) 3039.73 +Epoch -710 +---------------------------------- --------------- +2022-05-10 14:01:28.148260 PDT | [1] Epoch -709 finished +---------------------------------- --------------- +epoch -709 +replay_buffer/size 999033 +trainer/num train calls 292000 +trainer/Policy Loss -19.6251 +trainer/Log Pis Mean 23.8876 +trainer/Log Pis Std 12.5195 +trainer/Log Pis Max 70.8223 +trainer/Log Pis Min -6.67562 +trainer/policy/mean Mean -0.0349147 +trainer/policy/mean Std 0.90304 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83992 +trainer/policy/normal/std Std 0.617948 +trainer/policy/normal/std Max 5.47266 +trainer/policy/normal/std Min 0.259929 +trainer/policy/normal/log_std Mean 1.01029 +trainer/policy/normal/log_std Std 0.290217 +trainer/policy/normal/log_std Max 1.69977 +trainer/policy/normal/log_std Min -1.34735 +eval/num steps total 291303 +eval/num paths total 292 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.230894 +eval/Actions Std 0.890518 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67649 +time/logging (s) 0.00400504 +time/sampling batch (s) 0.278995 +time/saving (s) 0.00360304 +time/training (s) 6.93774 +time/epoch (s) 9.90084 +time/total (s) 3049.64 +Epoch -709 +---------------------------------- --------------- +2022-05-10 14:01:38.081005 PDT | [1] Epoch -708 finished +---------------------------------- --------------- +epoch -708 +replay_buffer/size 999033 +trainer/num train calls 293000 +trainer/Policy Loss -19.07 +trainer/Log Pis Mean 24.4382 +trainer/Log Pis Std 13.0983 +trainer/Log Pis Max 71.9504 +trainer/Log Pis Min -12.8888 +trainer/policy/mean Mean -0.0375183 +trainer/policy/mean Std 0.905278 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84533 +trainer/policy/normal/std Std 0.629661 +trainer/policy/normal/std Max 5.85716 +trainer/policy/normal/std Min 0.326625 +trainer/policy/normal/log_std Mean 1.01064 +trainer/policy/normal/log_std Std 0.299013 +trainer/policy/normal/log_std Max 1.76767 +trainer/policy/normal/log_std Min -1.11894 +eval/num steps total 292303 +eval/num paths total 293 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00590637 +eval/Actions Std 0.937192 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65554 +time/logging (s) 0.00382309 +time/sampling batch (s) 0.530702 +time/saving (s) 0.00348056 +time/training (s) 6.71752 +time/epoch (s) 9.91107 +time/total (s) 3059.55 +Epoch -708 +---------------------------------- --------------- +2022-05-10 14:01:48.452491 PDT | [1] Epoch -707 finished +---------------------------------- --------------- +epoch -707 +replay_buffer/size 999033 +trainer/num train calls 294000 +trainer/Policy Loss -19.8832 +trainer/Log Pis Mean 24.9428 +trainer/Log Pis Std 12.7267 +trainer/Log Pis Max 67.0932 +trainer/Log Pis Min -6.72867 +trainer/policy/mean Mean -0.0430333 +trainer/policy/mean Std 0.90781 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86025 +trainer/policy/normal/std Std 0.606039 +trainer/policy/normal/std Max 6.20195 +trainer/policy/normal/std Min 0.367178 +trainer/policy/normal/log_std Mean 1.02025 +trainer/policy/normal/log_std Std 0.275073 +trainer/policy/normal/log_std Max 1.82486 +trainer/policy/normal/log_std Min -1.00191 +eval/num steps total 293303 +eval/num paths total 294 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.30479 +eval/Actions Std 0.824539 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73616 +time/logging (s) 0.00385191 +time/sampling batch (s) 0.530125 +time/saving (s) 0.00358504 +time/training (s) 7.07652 +time/epoch (s) 10.3502 +time/total (s) 3069.9 +Epoch -707 +---------------------------------- --------------- +2022-05-10 14:01:58.527529 PDT | [1] Epoch -706 finished +---------------------------------- --------------- +epoch -706 +replay_buffer/size 999033 +trainer/num train calls 295000 +trainer/Policy Loss -20.3583 +trainer/Log Pis Mean 25.087 +trainer/Log Pis Std 12.5768 +trainer/Log Pis Max 69.3467 +trainer/Log Pis Min -7.80054 +trainer/policy/mean Mean -0.0390505 +trainer/policy/mean Std 0.906933 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78922 +trainer/policy/normal/std Std 0.615072 +trainer/policy/normal/std Max 6.62352 +trainer/policy/normal/std Min 0.34562 +trainer/policy/normal/log_std Mean 0.991552 +trainer/policy/normal/log_std Std 0.292364 +trainer/policy/normal/log_std Max 1.89063 +trainer/policy/normal/log_std Min -1.06242 +eval/num steps total 294303 +eval/num paths total 295 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.502067 +eval/Actions Std 0.768846 +eval/Actions Max 0.999993 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5386 +time/logging (s) 0.00378691 +time/sampling batch (s) 0.530096 +time/saving (s) 0.00343446 +time/training (s) 6.97756 +time/epoch (s) 10.0535 +time/total (s) 3079.96 +Epoch -706 +---------------------------------- --------------- +2022-05-10 14:02:09.476306 PDT | [1] Epoch -705 finished +---------------------------------- --------------- +epoch -705 +replay_buffer/size 999033 +trainer/num train calls 296000 +trainer/Policy Loss -19.7504 +trainer/Log Pis Mean 25.0584 +trainer/Log Pis Std 12.5098 +trainer/Log Pis Max 76.7322 +trainer/Log Pis Min -7.89635 +trainer/policy/mean Mean -0.029711 +trainer/policy/mean Std 0.903657 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84345 +trainer/policy/normal/std Std 0.614924 +trainer/policy/normal/std Max 5.32997 +trainer/policy/normal/std Min 0.328143 +trainer/policy/normal/log_std Mean 1.01091 +trainer/policy/normal/log_std Std 0.294957 +trainer/policy/normal/log_std Max 1.67335 +trainer/policy/normal/log_std Min -1.1143 +eval/num steps total 295303 +eval/num paths total 296 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0148612 +eval/Actions Std 0.905382 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78163 +time/logging (s) 0.00424346 +time/sampling batch (s) 1.02582 +time/saving (s) 0.0039772 +time/training (s) 7.11245 +time/epoch (s) 10.9281 +time/total (s) 3090.89 +Epoch -705 +---------------------------------- --------------- +2022-05-10 14:02:20.085170 PDT | [1] Epoch -704 finished +---------------------------------- --------------- +epoch -704 +replay_buffer/size 999033 +trainer/num train calls 297000 +trainer/Policy Loss -20.7844 +trainer/Log Pis Mean 25.4794 +trainer/Log Pis Std 13.1045 +trainer/Log Pis Max 71.2551 +trainer/Log Pis Min -7.67282 +trainer/policy/mean Mean -0.0490816 +trainer/policy/mean Std 0.908054 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85544 +trainer/policy/normal/std Std 0.609864 +trainer/policy/normal/std Max 5.95046 +trainer/policy/normal/std Min 0.285185 +trainer/policy/normal/log_std Mean 1.0161 +trainer/policy/normal/log_std Std 0.292293 +trainer/policy/normal/log_std Max 1.78347 +trainer/policy/normal/log_std Min -1.25462 +eval/num steps total 296303 +eval/num paths total 297 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130554 +eval/Actions Std 0.873861 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54086 +time/logging (s) 0.00499124 +time/sampling batch (s) 0.27846 +time/saving (s) 0.00560485 +time/training (s) 7.75808 +time/epoch (s) 10.588 +time/total (s) 3101.48 +Epoch -704 +---------------------------------- --------------- +2022-05-10 14:02:29.760465 PDT | [1] Epoch -703 finished +---------------------------------- --------------- +epoch -703 +replay_buffer/size 999033 +trainer/num train calls 298000 +trainer/Policy Loss -20.5276 +trainer/Log Pis Mean 25.722 +trainer/Log Pis Std 13.2142 +trainer/Log Pis Max 64.3593 +trainer/Log Pis Min -6.62787 +trainer/policy/mean Mean -0.0371083 +trainer/policy/mean Std 0.904733 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86115 +trainer/policy/normal/std Std 0.644364 +trainer/policy/normal/std Max 5.11511 +trainer/policy/normal/std Min 0.396057 +trainer/policy/normal/log_std Mean 1.0145 +trainer/policy/normal/log_std Std 0.305737 +trainer/policy/normal/log_std Max 1.6322 +trainer/policy/normal/log_std Min -0.926198 +eval/num steps total 297303 +eval/num paths total 298 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0751886 +eval/Actions Std 0.894077 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4711 +time/logging (s) 0.00366088 +time/sampling batch (s) 0.275418 +time/saving (s) 0.00344415 +time/training (s) 6.89941 +time/epoch (s) 9.65303 +time/total (s) 3111.14 +Epoch -703 +---------------------------------- --------------- +2022-05-10 14:02:40.758263 PDT | [1] Epoch -702 finished +---------------------------------- --------------- +epoch -702 +replay_buffer/size 999033 +trainer/num train calls 299000 +trainer/Policy Loss -20.1888 +trainer/Log Pis Mean 23.6092 +trainer/Log Pis Std 12.7252 +trainer/Log Pis Max 67.9657 +trainer/Log Pis Min -6.7793 +trainer/policy/mean Mean -0.0302433 +trainer/policy/mean Std 0.902215 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83871 +trainer/policy/normal/std Std 0.657585 +trainer/policy/normal/std Max 5.11891 +trainer/policy/normal/std Min 0.327742 +trainer/policy/normal/log_std Mean 1.0037 +trainer/policy/normal/log_std Std 0.318745 +trainer/policy/normal/log_std Max 1.63294 +trainer/policy/normal/log_std Min -1.11553 +eval/num steps total 298303 +eval/num paths total 299 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.140263 +eval/Actions Std 0.88521 +eval/Actions Max 0.999987 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57467 +time/logging (s) 0.00388862 +time/sampling batch (s) 0.274338 +time/saving (s) 0.00339115 +time/training (s) 8.12088 +time/epoch (s) 10.9772 +time/total (s) 3122.12 +Epoch -702 +---------------------------------- --------------- +2022-05-10 14:02:51.409978 PDT | [1] Epoch -701 finished +---------------------------------- --------------- +epoch -701 +replay_buffer/size 999033 +trainer/num train calls 300000 +trainer/Policy Loss -18.9321 +trainer/Log Pis Mean 23.8272 +trainer/Log Pis Std 13.4223 +trainer/Log Pis Max 63.7521 +trainer/Log Pis Min -11.8461 +trainer/policy/mean Mean -0.060678 +trainer/policy/mean Std 0.904611 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82755 +trainer/policy/normal/std Std 0.618362 +trainer/policy/normal/std Max 5.38878 +trainer/policy/normal/std Min 0.241514 +trainer/policy/normal/log_std Mean 1.00499 +trainer/policy/normal/log_std Std 0.296306 +trainer/policy/normal/log_std Max 1.68432 +trainer/policy/normal/log_std Min -1.42083 +eval/num steps total 299303 +eval/num paths total 300 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0212245 +eval/Actions Std 0.890186 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61838 +time/logging (s) 0.00369936 +time/sampling batch (s) 0.279163 +time/saving (s) 0.00630477 +time/training (s) 7.72312 +time/epoch (s) 10.6307 +time/total (s) 3132.75 +Epoch -701 +---------------------------------- --------------- +2022-05-10 14:03:01.808097 PDT | [1] Epoch -700 finished +---------------------------------- --------------- +epoch -700 +replay_buffer/size 999033 +trainer/num train calls 301000 +trainer/Policy Loss -19.0442 +trainer/Log Pis Mean 24.8613 +trainer/Log Pis Std 13.1445 +trainer/Log Pis Max 69.0935 +trainer/Log Pis Min -12.0596 +trainer/policy/mean Mean -0.0147529 +trainer/policy/mean Std 0.909768 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.94529 +trainer/policy/normal/std Std 0.659171 +trainer/policy/normal/std Max 5.66974 +trainer/policy/normal/std Min 0.311284 +trainer/policy/normal/log_std Mean 1.04382 +trainer/policy/normal/log_std Std 0.304961 +trainer/policy/normal/log_std Max 1.73514 +trainer/policy/normal/log_std Min -1.16705 +eval/num steps total 300303 +eval/num paths total 301 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0198192 +eval/Actions Std 0.90896 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44996 +time/logging (s) 0.00374945 +time/sampling batch (s) 0.275271 +time/saving (s) 0.00345174 +time/training (s) 7.64489 +time/epoch (s) 10.3773 +time/total (s) 3143.13 +Epoch -700 +---------------------------------- --------------- +2022-05-10 14:03:12.347548 PDT | [1] Epoch -699 finished +---------------------------------- --------------- +epoch -699 +replay_buffer/size 999033 +trainer/num train calls 302000 +trainer/Policy Loss -20.21 +trainer/Log Pis Mean 24.4684 +trainer/Log Pis Std 13.1214 +trainer/Log Pis Max 65.6071 +trainer/Log Pis Min -8.22115 +trainer/policy/mean Mean -0.0103053 +trainer/policy/mean Std 0.904143 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84955 +trainer/policy/normal/std Std 0.60499 +trainer/policy/normal/std Max 5.1145 +trainer/policy/normal/std Min 0.339117 +trainer/policy/normal/log_std Mean 1.01502 +trainer/policy/normal/log_std Std 0.28444 +trainer/policy/normal/log_std Max 1.63208 +trainer/policy/normal/log_std Min -1.08141 +eval/num steps total 301303 +eval/num paths total 302 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0803755 +eval/Actions Std 0.889831 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40692 +time/logging (s) 0.00411743 +time/sampling batch (s) 0.524808 +time/saving (s) 0.00387138 +time/training (s) 7.57925 +time/epoch (s) 10.519 +time/total (s) 3153.65 +Epoch -699 +---------------------------------- --------------- +2022-05-10 14:03:23.260937 PDT | [1] Epoch -698 finished +---------------------------------- --------------- +epoch -698 +replay_buffer/size 999033 +trainer/num train calls 303000 +trainer/Policy Loss -18.4879 +trainer/Log Pis Mean 23.9232 +trainer/Log Pis Std 13.9247 +trainer/Log Pis Max 67.7623 +trainer/Log Pis Min -8.29482 +trainer/policy/mean Mean -0.0374525 +trainer/policy/mean Std 0.904454 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.8328 +trainer/policy/normal/std Std 0.57124 +trainer/policy/normal/std Max 6.24725 +trainer/policy/normal/std Min 0.405179 +trainer/policy/normal/log_std Mean 1.01405 +trainer/policy/normal/log_std Std 0.257796 +trainer/policy/normal/log_std Max 1.83214 +trainer/policy/normal/log_std Min -0.903426 +eval/num steps total 302303 +eval/num paths total 303 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133292 +eval/Actions Std 0.916244 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58232 +time/logging (s) 0.00379903 +time/sampling batch (s) 0.529066 +time/saving (s) 0.00352763 +time/training (s) 7.773 +time/epoch (s) 10.8917 +time/total (s) 3164.55 +Epoch -698 +---------------------------------- --------------- +2022-05-10 14:03:34.144200 PDT | [1] Epoch -697 finished +---------------------------------- --------------- +epoch -697 +replay_buffer/size 999033 +trainer/num train calls 304000 +trainer/Policy Loss -18.3877 +trainer/Log Pis Mean 23.3719 +trainer/Log Pis Std 13.0798 +trainer/Log Pis Max 67.3714 +trainer/Log Pis Min -9.1411 +trainer/policy/mean Mean -0.0141562 +trainer/policy/mean Std 0.900989 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83422 +trainer/policy/normal/std Std 0.638949 +trainer/policy/normal/std Max 6.70141 +trainer/policy/normal/std Min 0.332252 +trainer/policy/normal/log_std Mean 1.00552 +trainer/policy/normal/log_std Std 0.302058 +trainer/policy/normal/log_std Max 1.90232 +trainer/policy/normal/log_std Min -1.10186 +eval/num steps total 303303 +eval/num paths total 304 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222592 +eval/Actions Std 0.915587 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68517 +time/logging (s) 0.00372376 +time/sampling batch (s) 0.279031 +time/saving (s) 0.00345142 +time/training (s) 7.89056 +time/epoch (s) 10.8619 +time/total (s) 3175.41 +Epoch -697 +---------------------------------- --------------- +2022-05-10 14:03:44.609211 PDT | [1] Epoch -696 finished +---------------------------------- --------------- +epoch -696 +replay_buffer/size 999033 +trainer/num train calls 305000 +trainer/Policy Loss -19.5328 +trainer/Log Pis Mean 24.0043 +trainer/Log Pis Std 12.7436 +trainer/Log Pis Max 68.0925 +trainer/Log Pis Min -5.10838 +trainer/policy/mean Mean -0.0518747 +trainer/policy/mean Std 0.903627 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77679 +trainer/policy/normal/std Std 0.599531 +trainer/policy/normal/std Max 5.52547 +trainer/policy/normal/std Min 0.43138 +trainer/policy/normal/log_std Mean 0.989392 +trainer/policy/normal/log_std Std 0.279556 +trainer/policy/normal/log_std Max 1.70937 +trainer/policy/normal/log_std Min -0.840767 +eval/num steps total 304303 +eval/num paths total 305 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0628892 +eval/Actions Std 0.938258 +eval/Actions Max 0.999986 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55162 +time/logging (s) 0.00374562 +time/sampling batch (s) 0.276064 +time/saving (s) 0.0035112 +time/training (s) 7.60904 +time/epoch (s) 10.444 +time/total (s) 3185.86 +Epoch -696 +---------------------------------- --------------- +2022-05-10 14:03:55.052326 PDT | [1] Epoch -695 finished +---------------------------------- --------------- +epoch -695 +replay_buffer/size 999033 +trainer/num train calls 306000 +trainer/Policy Loss -18.3083 +trainer/Log Pis Mean 22.5782 +trainer/Log Pis Std 12.6147 +trainer/Log Pis Max 64.5709 +trainer/Log Pis Min -10.4214 +trainer/policy/mean Mean -0.00478436 +trainer/policy/mean Std 0.903909 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75016 +trainer/policy/normal/std Std 0.600661 +trainer/policy/normal/std Max 6.00537 +trainer/policy/normal/std Min 0.37004 +trainer/policy/normal/log_std Mean 0.977865 +trainer/policy/normal/log_std Std 0.291243 +trainer/policy/normal/log_std Max 1.79265 +trainer/policy/normal/log_std Min -0.994145 +eval/num steps total 305303 +eval/num paths total 306 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00766753 +eval/Actions Std 0.978206 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53152 +time/logging (s) 0.00373169 +time/sampling batch (s) 0.525696 +time/saving (s) 0.00341263 +time/training (s) 7.35764 +time/epoch (s) 10.422 +time/total (s) 3196.29 +Epoch -695 +---------------------------------- --------------- +2022-05-10 14:04:05.845584 PDT | [1] Epoch -694 finished +---------------------------------- --------------- +epoch -694 +replay_buffer/size 999033 +trainer/num train calls 307000 +trainer/Policy Loss -19.2928 +trainer/Log Pis Mean 25.58 +trainer/Log Pis Std 12.9772 +trainer/Log Pis Max 64.2687 +trainer/Log Pis Min -13.0963 +trainer/policy/mean Mean -0.0329017 +trainer/policy/mean Std 0.904365 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83532 +trainer/policy/normal/std Std 0.6377 +trainer/policy/normal/std Max 6.06517 +trainer/policy/normal/std Min 0.306054 +trainer/policy/normal/log_std Mean 1.00506 +trainer/policy/normal/log_std Std 0.309045 +trainer/policy/normal/log_std Max 1.80256 +trainer/policy/normal/log_std Min -1.18399 +eval/num steps total 306303 +eval/num paths total 307 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178782 +eval/Actions Std 0.905135 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46756 +time/logging (s) 0.00431847 +time/sampling batch (s) 0.777258 +time/saving (s) 0.00414249 +time/training (s) 7.51944 +time/epoch (s) 10.7727 +time/total (s) 3207.06 +Epoch -694 +---------------------------------- --------------- +2022-05-10 14:04:16.046897 PDT | [1] Epoch -693 finished +---------------------------------- --------------- +epoch -693 +replay_buffer/size 999033 +trainer/num train calls 308000 +trainer/Policy Loss -19.3684 +trainer/Log Pis Mean 25.1292 +trainer/Log Pis Std 14.7786 +trainer/Log Pis Max 72.8709 +trainer/Log Pis Min -8.22617 +trainer/policy/mean Mean -0.0507297 +trainer/policy/mean Std 0.906026 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84034 +trainer/policy/normal/std Std 0.601997 +trainer/policy/normal/std Max 6.05133 +trainer/policy/normal/std Min 0.356766 +trainer/policy/normal/log_std Mean 1.01258 +trainer/policy/normal/log_std Std 0.279749 +trainer/policy/normal/log_std Max 1.80028 +trainer/policy/normal/log_std Min -1.03067 +eval/num steps total 307303 +eval/num paths total 308 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.166695 +eval/Actions Std 0.854807 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63494 +time/logging (s) 0.00381352 +time/sampling batch (s) 0.275548 +time/saving (s) 0.00360373 +time/training (s) 7.26139 +time/epoch (s) 10.1793 +time/total (s) 3217.24 +Epoch -693 +---------------------------------- --------------- +2022-05-10 14:04:27.546105 PDT | [1] Epoch -692 finished +---------------------------------- --------------- +epoch -692 +replay_buffer/size 999033 +trainer/num train calls 309000 +trainer/Policy Loss -20.4133 +trainer/Log Pis Mean 25.4931 +trainer/Log Pis Std 13.6831 +trainer/Log Pis Max 82.932 +trainer/Log Pis Min -10.5596 +trainer/policy/mean Mean -0.0510271 +trainer/policy/mean Std 0.908395 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87617 +trainer/policy/normal/std Std 0.608429 +trainer/policy/normal/std Max 5.42316 +trainer/policy/normal/std Min 0.34426 +trainer/policy/normal/log_std Mean 1.02624 +trainer/policy/normal/log_std Std 0.271907 +trainer/policy/normal/log_std Max 1.69068 +trainer/policy/normal/log_std Min -1.06636 +eval/num steps total 308303 +eval/num paths total 309 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00695326 +eval/Actions Std 0.923021 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53046 +time/logging (s) 0.00372214 +time/sampling batch (s) 1.03573 +time/saving (s) 0.00345496 +time/training (s) 7.90405 +time/epoch (s) 11.4774 +time/total (s) 3228.73 +Epoch -692 +---------------------------------- --------------- +2022-05-10 14:04:37.397825 PDT | [1] Epoch -691 finished +---------------------------------- --------------- +epoch -691 +replay_buffer/size 999033 +trainer/num train calls 310000 +trainer/Policy Loss -19.6286 +trainer/Log Pis Mean 23.8427 +trainer/Log Pis Std 12.6694 +trainer/Log Pis Max 64.2063 +trainer/Log Pis Min -9.99921 +trainer/policy/mean Mean -0.0324462 +trainer/policy/mean Std 0.9031 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81019 +trainer/policy/normal/std Std 0.611882 +trainer/policy/normal/std Max 5.50324 +trainer/policy/normal/std Min 0.39614 +trainer/policy/normal/log_std Mean 1.00065 +trainer/policy/normal/log_std Std 0.283923 +trainer/policy/normal/log_std Max 1.70534 +trainer/policy/normal/log_std Min -0.925987 +eval/num steps total 309303 +eval/num paths total 310 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0775238 +eval/Actions Std 0.914027 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44461 +time/logging (s) 0.00379675 +time/sampling batch (s) 0.280346 +time/saving (s) 0.00343883 +time/training (s) 7.09799 +time/epoch (s) 9.83019 +time/total (s) 3238.56 +Epoch -691 +---------------------------------- --------------- +2022-05-10 14:04:48.224082 PDT | [1] Epoch -690 finished +---------------------------------- --------------- +epoch -690 +replay_buffer/size 999033 +trainer/num train calls 311000 +trainer/Policy Loss -19.2803 +trainer/Log Pis Mean 24.7708 +trainer/Log Pis Std 13.167 +trainer/Log Pis Max 68.2872 +trainer/Log Pis Min -6.9925 +trainer/policy/mean Mean -0.0380747 +trainer/policy/mean Std 0.906971 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85464 +trainer/policy/normal/std Std 0.633693 +trainer/policy/normal/std Max 6.09728 +trainer/policy/normal/std Min 0.371663 +trainer/policy/normal/log_std Mean 1.0127 +trainer/policy/normal/log_std Std 0.305329 +trainer/policy/normal/log_std Max 1.80784 +trainer/policy/normal/log_std Min -0.989767 +eval/num steps total 310303 +eval/num paths total 311 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0715215 +eval/Actions Std 0.916486 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6137 +time/logging (s) 0.00379364 +time/sampling batch (s) 0.537287 +time/saving (s) 0.00346555 +time/training (s) 7.64628 +time/epoch (s) 10.8045 +time/total (s) 3249.37 +Epoch -690 +---------------------------------- --------------- +2022-05-10 14:04:57.704928 PDT | [1] Epoch -689 finished +---------------------------------- --------------- +epoch -689 +replay_buffer/size 999033 +trainer/num train calls 312000 +trainer/Policy Loss -19.0729 +trainer/Log Pis Mean 24.6885 +trainer/Log Pis Std 13.3743 +trainer/Log Pis Max 68.5881 +trainer/Log Pis Min -8.41855 +trainer/policy/mean Mean -0.0447635 +trainer/policy/mean Std 0.904806 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82744 +trainer/policy/normal/std Std 0.62321 +trainer/policy/normal/std Max 5.36487 +trainer/policy/normal/std Min 0.373653 +trainer/policy/normal/log_std Mean 1.00428 +trainer/policy/normal/log_std Std 0.298161 +trainer/policy/normal/log_std Max 1.67987 +trainer/policy/normal/log_std Min -0.984428 +eval/num steps total 311303 +eval/num paths total 312 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0380197 +eval/Actions Std 0.90751 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46383 +time/logging (s) 0.0038373 +time/sampling batch (s) 0.281426 +time/saving (s) 0.00345918 +time/training (s) 6.70682 +time/epoch (s) 9.45938 +time/total (s) 3258.83 +Epoch -689 +---------------------------------- --------------- +2022-05-10 14:05:08.654329 PDT | [1] Epoch -688 finished +---------------------------------- --------------- +epoch -688 +replay_buffer/size 999033 +trainer/num train calls 313000 +trainer/Policy Loss -19.7293 +trainer/Log Pis Mean 23.9625 +trainer/Log Pis Std 12.6668 +trainer/Log Pis Max 68.6353 +trainer/Log Pis Min -3.75651 +trainer/policy/mean Mean -0.044559 +trainer/policy/mean Std 0.905042 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.77832 +trainer/policy/normal/std Std 0.614345 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.358296 +trainer/policy/normal/log_std Mean 0.987929 +trainer/policy/normal/log_std Std 0.291092 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.0264 +eval/num steps total 312303 +eval/num paths total 313 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109585 +eval/Actions Std 0.908497 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68723 +time/logging (s) 0.0041153 +time/sampling batch (s) 0.532891 +time/saving (s) 0.00384313 +time/training (s) 7.69982 +time/epoch (s) 10.9279 +time/total (s) 3269.76 +Epoch -688 +---------------------------------- --------------- +2022-05-10 14:05:19.518486 PDT | [1] Epoch -687 finished +---------------------------------- --------------- +epoch -687 +replay_buffer/size 999033 +trainer/num train calls 314000 +trainer/Policy Loss -19.7708 +trainer/Log Pis Mean 25.3392 +trainer/Log Pis Std 13.1423 +trainer/Log Pis Max 61.235 +trainer/Log Pis Min -7.03863 +trainer/policy/mean Mean -0.0502143 +trainer/policy/mean Std 0.906443 +trainer/policy/mean Max 0.999968 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.86991 +trainer/policy/normal/std Std 0.627211 +trainer/policy/normal/std Max 6.02017 +trainer/policy/normal/std Min 0.359579 +trainer/policy/normal/log_std Mean 1.02029 +trainer/policy/normal/log_std Std 0.293364 +trainer/policy/normal/log_std Max 1.79512 +trainer/policy/normal/log_std Min -1.02282 +eval/num steps total 313303 +eval/num paths total 314 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105542 +eval/Actions Std 0.898478 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65306 +time/logging (s) 0.00384912 +time/sampling batch (s) 0.282341 +time/saving (s) 0.00355786 +time/training (s) 7.89918 +time/epoch (s) 10.842 +time/total (s) 3280.6 +Epoch -687 +---------------------------------- --------------- +2022-05-10 14:05:30.094841 PDT | [1] Epoch -686 finished +---------------------------------- --------------- +epoch -686 +replay_buffer/size 999033 +trainer/num train calls 315000 +trainer/Policy Loss -20.1749 +trainer/Log Pis Mean 25.6262 +trainer/Log Pis Std 13.1432 +trainer/Log Pis Max 67.1112 +trainer/Log Pis Min -4.02117 +trainer/policy/mean Mean -0.0395182 +trainer/policy/mean Std 0.906843 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.88326 +trainer/policy/normal/std Std 0.626964 +trainer/policy/normal/std Max 5.7174 +trainer/policy/normal/std Min 0.329798 +trainer/policy/normal/log_std Mean 1.02503 +trainer/policy/normal/log_std Std 0.2928 +trainer/policy/normal/log_std Max 1.74351 +trainer/policy/normal/log_std Min -1.10928 +eval/num steps total 314303 +eval/num paths total 315 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0540506 +eval/Actions Std 0.913565 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58941 +time/logging (s) 0.00369931 +time/sampling batch (s) 0.532047 +time/saving (s) 0.00340986 +time/training (s) 7.4259 +time/epoch (s) 10.5545 +time/total (s) 3291.16 +Epoch -686 +---------------------------------- --------------- +2022-05-10 14:05:40.884453 PDT | [1] Epoch -685 finished +---------------------------------- --------------- +epoch -685 +replay_buffer/size 999033 +trainer/num train calls 316000 +trainer/Policy Loss -18.7259 +trainer/Log Pis Mean 24.3934 +trainer/Log Pis Std 14.1846 +trainer/Log Pis Max 105.598 +trainer/Log Pis Min -9.99232 +trainer/policy/mean Mean -0.0351328 +trainer/policy/mean Std 0.900916 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87204 +trainer/policy/normal/std Std 0.628245 +trainer/policy/normal/std Max 6.09737 +trainer/policy/normal/std Min 0.35027 +trainer/policy/normal/log_std Mean 1.02237 +trainer/policy/normal/log_std Std 0.284474 +trainer/policy/normal/log_std Max 1.80786 +trainer/policy/normal/log_std Min -1.04905 +eval/num steps total 315303 +eval/num paths total 316 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.006536 +eval/Actions Std 0.901316 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68448 +time/logging (s) 0.00387062 +time/sampling batch (s) 0.52819 +time/saving (s) 0.00345156 +time/training (s) 7.5485 +time/epoch (s) 10.7685 +time/total (s) 3301.93 +Epoch -685 +---------------------------------- --------------- +2022-05-10 14:05:50.992990 PDT | [1] Epoch -684 finished +---------------------------------- --------------- +epoch -684 +replay_buffer/size 999033 +trainer/num train calls 317000 +trainer/Policy Loss -19.6287 +trainer/Log Pis Mean 24.3964 +trainer/Log Pis Std 14.3252 +trainer/Log Pis Max 63.8232 +trainer/Log Pis Min -13.3815 +trainer/policy/mean Mean -0.0515408 +trainer/policy/mean Std 0.907582 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88279 +trainer/policy/normal/std Std 0.629674 +trainer/policy/normal/std Max 5.93032 +trainer/policy/normal/std Min 0.320727 +trainer/policy/normal/log_std Mean 1.02479 +trainer/policy/normal/log_std Std 0.292693 +trainer/policy/normal/log_std Max 1.78008 +trainer/policy/normal/log_std Min -1.13717 +eval/num steps total 316303 +eval/num paths total 317 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.208899 +eval/Actions Std 0.878258 +eval/Actions Max 0.999985 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66385 +time/logging (s) 0.00378981 +time/sampling batch (s) 0.528951 +time/saving (s) 0.00340291 +time/training (s) 6.88715 +time/epoch (s) 10.0871 +time/total (s) 3312.02 +Epoch -684 +---------------------------------- --------------- +2022-05-10 14:06:01.260388 PDT | [1] Epoch -683 finished +---------------------------------- --------------- +epoch -683 +replay_buffer/size 999033 +trainer/num train calls 318000 +trainer/Policy Loss -18.7 +trainer/Log Pis Mean 24.9167 +trainer/Log Pis Std 13.8606 +trainer/Log Pis Max 69.7519 +trainer/Log Pis Min -8.22457 +trainer/policy/mean Mean -0.0380401 +trainer/policy/mean Std 0.909617 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78752 +trainer/policy/normal/std Std 0.600935 +trainer/policy/normal/std Max 5.19394 +trainer/policy/normal/std Min 0.288786 +trainer/policy/normal/log_std Mean 0.992652 +trainer/policy/normal/log_std Std 0.285311 +trainer/policy/normal/log_std Max 1.64749 +trainer/policy/normal/log_std Min -1.24207 +eval/num steps total 317303 +eval/num paths total 318 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.325935 +eval/Actions Std 0.865909 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63089 +time/logging (s) 0.00438194 +time/sampling batch (s) 0.527739 +time/saving (s) 0.00394097 +time/training (s) 7.07951 +time/epoch (s) 10.2465 +time/total (s) 3322.27 +Epoch -683 +---------------------------------- --------------- +2022-05-10 14:06:10.730191 PDT | [1] Epoch -682 finished +---------------------------------- --------------- +epoch -682 +replay_buffer/size 999033 +trainer/num train calls 319000 +trainer/Policy Loss -20.0419 +trainer/Log Pis Mean 25.3331 +trainer/Log Pis Std 13.1526 +trainer/Log Pis Max 62.7605 +trainer/Log Pis Min -9.04097 +trainer/policy/mean Mean -0.049025 +trainer/policy/mean Std 0.906739 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81336 +trainer/policy/normal/std Std 0.624018 +trainer/policy/normal/std Max 5.20445 +trainer/policy/normal/std Min 0.288272 +trainer/policy/normal/log_std Mean 0.999671 +trainer/policy/normal/log_std Std 0.295162 +trainer/policy/normal/log_std Max 1.64951 +trainer/policy/normal/log_std Min -1.24385 +eval/num steps total 318303 +eval/num paths total 319 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.36023 +eval/Actions Std 0.874184 +eval/Actions Max 0.999977 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59904 +time/logging (s) 0.00386843 +time/sampling batch (s) 0.275285 +time/saving (s) 0.00357517 +time/training (s) 6.56588 +time/epoch (s) 9.44765 +time/total (s) 3331.72 +Epoch -682 +---------------------------------- --------------- +2022-05-10 14:06:21.355902 PDT | [1] Epoch -681 finished +---------------------------------- --------------- +epoch -681 +replay_buffer/size 999033 +trainer/num train calls 320000 +trainer/Policy Loss -20.2562 +trainer/Log Pis Mean 24.0179 +trainer/Log Pis Std 13.0958 +trainer/Log Pis Max 63.2086 +trainer/Log Pis Min -6.95581 +trainer/policy/mean Mean -0.0205743 +trainer/policy/mean Std 0.902505 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78663 +trainer/policy/normal/std Std 0.618143 +trainer/policy/normal/std Max 5.62949 +trainer/policy/normal/std Min 0.385516 +trainer/policy/normal/log_std Mean 0.989857 +trainer/policy/normal/log_std Std 0.296693 +trainer/policy/normal/log_std Max 1.72802 +trainer/policy/normal/log_std Min -0.953172 +eval/num steps total 319303 +eval/num paths total 320 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.127947 +eval/Actions Std 0.798521 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60521 +time/logging (s) 0.00384021 +time/sampling batch (s) 0.27705 +time/saving (s) 0.00343753 +time/training (s) 7.71463 +time/epoch (s) 10.6042 +time/total (s) 3342.33 +Epoch -681 +---------------------------------- --------------- +2022-05-10 14:06:31.119630 PDT | [1] Epoch -680 finished +---------------------------------- -------------- +epoch -680 +replay_buffer/size 999033 +trainer/num train calls 321000 +trainer/Policy Loss -19.6198 +trainer/Log Pis Mean 25.4466 +trainer/Log Pis Std 13.1859 +trainer/Log Pis Max 72.6567 +trainer/Log Pis Min -6.89214 +trainer/policy/mean Mean -0.0209472 +trainer/policy/mean Std 0.909892 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8779 +trainer/policy/normal/std Std 0.638196 +trainer/policy/normal/std Max 5.65305 +trainer/policy/normal/std Min 0.316306 +trainer/policy/normal/log_std Mean 1.01993 +trainer/policy/normal/log_std Std 0.312067 +trainer/policy/normal/log_std Max 1.7322 +trainer/policy/normal/log_std Min -1.15105 +eval/num steps total 320303 +eval/num paths total 321 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.211939 +eval/Actions Std 0.930171 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79883 +time/logging (s) 0.0039099 +time/sampling batch (s) 0.275935 +time/saving (s) 0.0034282 +time/training (s) 6.66053 +time/epoch (s) 9.74264 +time/total (s) 3352.08 +Epoch -680 +---------------------------------- -------------- +2022-05-10 14:06:40.673242 PDT | [1] Epoch -679 finished +---------------------------------- --------------- +epoch -679 +replay_buffer/size 999033 +trainer/num train calls 322000 +trainer/Policy Loss -19.0918 +trainer/Log Pis Mean 24.7276 +trainer/Log Pis Std 13.4779 +trainer/Log Pis Max 65.8221 +trainer/Log Pis Min -8.50723 +trainer/policy/mean Mean -0.0568929 +trainer/policy/mean Std 0.910114 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82841 +trainer/policy/normal/std Std 0.611545 +trainer/policy/normal/std Max 4.94702 +trainer/policy/normal/std Min 0.412485 +trainer/policy/normal/log_std Mean 1.00633 +trainer/policy/normal/log_std Std 0.28974 +trainer/policy/normal/log_std Max 1.59879 +trainer/policy/normal/log_std Min -0.885556 +eval/num steps total 321303 +eval/num paths total 322 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.332594 +eval/Actions Std 0.817666 +eval/Actions Max 0.999994 +eval/Actions Min -0.999948 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48438 +time/logging (s) 0.00448168 +time/sampling batch (s) 0.543248 +time/saving (s) 0.00472651 +time/training (s) 6.49556 +time/epoch (s) 9.5324 +time/total (s) 3361.61 +Epoch -679 +---------------------------------- --------------- +2022-05-10 14:06:51.647254 PDT | [1] Epoch -678 finished +---------------------------------- --------------- +epoch -678 +replay_buffer/size 999033 +trainer/num train calls 323000 +trainer/Policy Loss -20.5748 +trainer/Log Pis Mean 23.2604 +trainer/Log Pis Std 13.1307 +trainer/Log Pis Max 64.9744 +trainer/Log Pis Min -9.41527 +trainer/policy/mean Mean -0.0511502 +trainer/policy/mean Std 0.905032 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84996 +trainer/policy/normal/std Std 0.619448 +trainer/policy/normal/std Max 5.75497 +trainer/policy/normal/std Min 0.351637 +trainer/policy/normal/log_std Mean 1.01431 +trainer/policy/normal/log_std Std 0.287234 +trainer/policy/normal/log_std Max 1.75006 +trainer/policy/normal/log_std Min -1.04516 +eval/num steps total 322303 +eval/num paths total 323 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.192133 +eval/Actions Std 0.937608 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76315 +time/logging (s) 0.00438816 +time/sampling batch (s) 0.542151 +time/saving (s) 0.00464582 +time/training (s) 7.63665 +time/epoch (s) 10.951 +time/total (s) 3372.57 +Epoch -678 +---------------------------------- --------------- +2022-05-10 14:07:01.648979 PDT | [1] Epoch -677 finished +---------------------------------- --------------- +epoch -677 +replay_buffer/size 999033 +trainer/num train calls 324000 +trainer/Policy Loss -19.8478 +trainer/Log Pis Mean 25.2477 +trainer/Log Pis Std 13.1508 +trainer/Log Pis Max 66.2399 +trainer/Log Pis Min -3.47465 +trainer/policy/mean Mean -0.0326743 +trainer/policy/mean Std 0.907769 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84999 +trainer/policy/normal/std Std 0.61979 +trainer/policy/normal/std Max 6.67259 +trainer/policy/normal/std Min 0.385937 +trainer/policy/normal/log_std Mean 1.01366 +trainer/policy/normal/log_std Std 0.291285 +trainer/policy/normal/log_std Max 1.89801 +trainer/policy/normal/log_std Min -0.95208 +eval/num steps total 323303 +eval/num paths total 324 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139154 +eval/Actions Std 0.884543 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63601 +time/logging (s) 0.00455812 +time/sampling batch (s) 0.52864 +time/saving (s) 0.00442207 +time/training (s) 6.80592 +time/epoch (s) 9.97954 +time/total (s) 3382.55 +Epoch -677 +---------------------------------- --------------- +2022-05-10 14:07:12.887199 PDT | [1] Epoch -676 finished +---------------------------------- --------------- +epoch -676 +replay_buffer/size 999033 +trainer/num train calls 325000 +trainer/Policy Loss -18.7067 +trainer/Log Pis Mean 24.0127 +trainer/Log Pis Std 14.1499 +trainer/Log Pis Max 72.543 +trainer/Log Pis Min -10.9246 +trainer/policy/mean Mean -0.0382236 +trainer/policy/mean Std 0.90336 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.75877 +trainer/policy/normal/std Std 0.585669 +trainer/policy/normal/std Max 5.18087 +trainer/policy/normal/std Min 0.290757 +trainer/policy/normal/log_std Mean 0.982999 +trainer/policy/normal/log_std Std 0.28217 +trainer/policy/normal/log_std Max 1.64497 +trainer/policy/normal/log_std Min -1.23527 +eval/num steps total 324303 +eval/num paths total 325 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0997725 +eval/Actions Std 0.870238 +eval/Actions Max 0.999988 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60247 +time/logging (s) 0.00369126 +time/sampling batch (s) 0.526664 +time/saving (s) 0.00350048 +time/training (s) 8.07907 +time/epoch (s) 11.2154 +time/total (s) 3393.77 +Epoch -676 +---------------------------------- --------------- +2022-05-10 14:07:23.848129 PDT | [1] Epoch -675 finished +---------------------------------- --------------- +epoch -675 +replay_buffer/size 999033 +trainer/num train calls 326000 +trainer/Policy Loss -21.575 +trainer/Log Pis Mean 25.2357 +trainer/Log Pis Std 13.0495 +trainer/Log Pis Max 74.9344 +trainer/Log Pis Min -5.32301 +trainer/policy/mean Mean -0.0697392 +trainer/policy/mean Std 0.906834 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85112 +trainer/policy/normal/std Std 0.622134 +trainer/policy/normal/std Max 5.25443 +trainer/policy/normal/std Min 0.331256 +trainer/policy/normal/log_std Mean 1.01434 +trainer/policy/normal/log_std Std 0.288831 +trainer/policy/normal/log_std Max 1.65907 +trainer/policy/normal/log_std Min -1.10486 +eval/num steps total 325303 +eval/num paths total 326 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0907658 +eval/Actions Std 0.880834 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56391 +time/logging (s) 0.0041062 +time/sampling batch (s) 0.527265 +time/saving (s) 0.00391434 +time/training (s) 7.84077 +time/epoch (s) 10.94 +time/total (s) 3404.71 +Epoch -675 +---------------------------------- --------------- +2022-05-10 14:07:34.463434 PDT | [1] Epoch -674 finished +---------------------------------- --------------- +epoch -674 +replay_buffer/size 999033 +trainer/num train calls 327000 +trainer/Policy Loss -20.2509 +trainer/Log Pis Mean 25.0945 +trainer/Log Pis Std 13.6836 +trainer/Log Pis Max 75.0704 +trainer/Log Pis Min -10.1121 +trainer/policy/mean Mean -0.0783856 +trainer/policy/mean Std 0.907474 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79616 +trainer/policy/normal/std Std 0.614605 +trainer/policy/normal/std Max 5.23127 +trainer/policy/normal/std Min 0.32039 +trainer/policy/normal/log_std Mean 0.992841 +trainer/policy/normal/log_std Std 0.301106 +trainer/policy/normal/log_std Max 1.65465 +trainer/policy/normal/log_std Min -1.13822 +eval/num steps total 325752 +eval/num paths total 327 +eval/path length Mean 449 +eval/path length Std 0 +eval/path length Max 449 +eval/path length Min 449 +eval/Rewards Mean 0.00222717 +eval/Rewards Std 0.0471403 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0127316 +eval/Actions Std 0.90139 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.56856 +time/logging (s) 0.00226151 +time/sampling batch (s) 0.274549 +time/saving (s) 0.00342917 +time/training (s) 7.74324 +time/epoch (s) 10.592 +time/total (s) 3415.31 +Epoch -674 +---------------------------------- --------------- +2022-05-10 14:07:44.388233 PDT | [1] Epoch -673 finished +---------------------------------- --------------- +epoch -673 +replay_buffer/size 999033 +trainer/num train calls 328000 +trainer/Policy Loss -20.6004 +trainer/Log Pis Mean 24.1494 +trainer/Log Pis Std 13.4068 +trainer/Log Pis Max 67.8983 +trainer/Log Pis Min -6.68473 +trainer/policy/mean Mean -0.0412944 +trainer/policy/mean Std 0.905004 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81818 +trainer/policy/normal/std Std 0.59735 +trainer/policy/normal/std Max 5.59285 +trainer/policy/normal/std Min 0.346165 +trainer/policy/normal/log_std Mean 1.00338 +trainer/policy/normal/log_std Std 0.288581 +trainer/policy/normal/log_std Max 1.72149 +trainer/policy/normal/log_std Min -1.06084 +eval/num steps total 326752 +eval/num paths total 328 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0548366 +eval/Actions Std 0.917157 +eval/Actions Max 0.999983 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71759 +time/logging (s) 0.00422755 +time/sampling batch (s) 0.279993 +time/saving (s) 0.00389163 +time/training (s) 6.89967 +time/epoch (s) 9.90537 +time/total (s) 3425.22 +Epoch -673 +---------------------------------- --------------- +2022-05-10 14:07:53.792442 PDT | [1] Epoch -672 finished +---------------------------------- --------------- +epoch -672 +replay_buffer/size 999033 +trainer/num train calls 329000 +trainer/Policy Loss -18.9143 +trainer/Log Pis Mean 22.9545 +trainer/Log Pis Std 13.2706 +trainer/Log Pis Max 68.1125 +trainer/Log Pis Min -8.33048 +trainer/policy/mean Mean -0.0262992 +trainer/policy/mean Std 0.905507 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78655 +trainer/policy/normal/std Std 0.602462 +trainer/policy/normal/std Max 5.65734 +trainer/policy/normal/std Min 0.30093 +trainer/policy/normal/log_std Mean 0.991516 +trainer/policy/normal/log_std Std 0.290261 +trainer/policy/normal/log_std Max 1.73295 +trainer/policy/normal/log_std Min -1.20088 +eval/num steps total 327752 +eval/num paths total 329 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0179772 +eval/Actions Std 0.91441 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39322 +time/logging (s) 0.00381044 +time/sampling batch (s) 0.27901 +time/saving (s) 0.00367533 +time/training (s) 6.70184 +time/epoch (s) 9.38156 +time/total (s) 3434.6 +Epoch -672 +---------------------------------- --------------- +2022-05-10 14:08:03.920216 PDT | [1] Epoch -671 finished +---------------------------------- --------------- +epoch -671 +replay_buffer/size 999033 +trainer/num train calls 330000 +trainer/Policy Loss -19.3265 +trainer/Log Pis Mean 24.3481 +trainer/Log Pis Std 13.1849 +trainer/Log Pis Max 69.2036 +trainer/Log Pis Min -6.38022 +trainer/policy/mean Mean -0.0263104 +trainer/policy/mean Std 0.903535 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80907 +trainer/policy/normal/std Std 0.618249 +trainer/policy/normal/std Max 5.93974 +trainer/policy/normal/std Min 0.291115 +trainer/policy/normal/log_std Mean 0.998739 +trainer/policy/normal/log_std Std 0.291992 +trainer/policy/normal/log_std Max 1.78167 +trainer/policy/normal/log_std Min -1.23404 +eval/num steps total 328752 +eval/num paths total 330 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0202978 +eval/Actions Std 0.908668 +eval/Actions Max 0.999996 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64585 +time/logging (s) 0.00378213 +time/sampling batch (s) 0.279529 +time/saving (s) 0.00346967 +time/training (s) 7.17342 +time/epoch (s) 10.1061 +time/total (s) 3444.71 +Epoch -671 +---------------------------------- --------------- +2022-05-10 14:08:14.485815 PDT | [1] Epoch -670 finished +---------------------------------- --------------- +epoch -670 +replay_buffer/size 999033 +trainer/num train calls 331000 +trainer/Policy Loss -18.4685 +trainer/Log Pis Mean 25.0012 +trainer/Log Pis Std 13.5853 +trainer/Log Pis Max 74.9692 +trainer/Log Pis Min -11.405 +trainer/policy/mean Mean -0.0192509 +trainer/policy/mean Std 0.906951 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85937 +trainer/policy/normal/std Std 0.605647 +trainer/policy/normal/std Max 6.303 +trainer/policy/normal/std Min 0.371258 +trainer/policy/normal/log_std Mean 1.01923 +trainer/policy/normal/log_std Std 0.280184 +trainer/policy/normal/log_std Max 1.84103 +trainer/policy/normal/log_std Min -0.990859 +eval/num steps total 329752 +eval/num paths total 331 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112502 +eval/Actions Std 0.897916 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68167 +time/logging (s) 0.0037021 +time/sampling batch (s) 0.279917 +time/saving (s) 0.00343717 +time/training (s) 7.57528 +time/epoch (s) 10.544 +time/total (s) 3455.26 +Epoch -670 +---------------------------------- --------------- +2022-05-10 14:08:25.223395 PDT | [1] Epoch -669 finished +---------------------------------- --------------- +epoch -669 +replay_buffer/size 999033 +trainer/num train calls 332000 +trainer/Policy Loss -19.963 +trainer/Log Pis Mean 25.0585 +trainer/Log Pis Std 13.5635 +trainer/Log Pis Max 73.1607 +trainer/Log Pis Min -10.0724 +trainer/policy/mean Mean -0.0357765 +trainer/policy/mean Std 0.908568 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81919 +trainer/policy/normal/std Std 0.614116 +trainer/policy/normal/std Max 5.38653 +trainer/policy/normal/std Min 0.343427 +trainer/policy/normal/log_std Mean 1.00184 +trainer/policy/normal/log_std Std 0.297757 +trainer/policy/normal/log_std Max 1.6839 +trainer/policy/normal/log_std Min -1.06878 +eval/num steps total 330752 +eval/num paths total 332 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0910406 +eval/Actions Std 0.915111 +eval/Actions Max 0.999996 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53703 +time/logging (s) 0.00391018 +time/sampling batch (s) 0.276434 +time/saving (s) 0.0034351 +time/training (s) 7.89567 +time/epoch (s) 10.7165 +time/total (s) 3465.98 +Epoch -669 +---------------------------------- --------------- +2022-05-10 14:08:35.027825 PDT | [1] Epoch -668 finished +---------------------------------- --------------- +epoch -668 +replay_buffer/size 999033 +trainer/num train calls 333000 +trainer/Policy Loss -19.4415 +trainer/Log Pis Mean 25.8125 +trainer/Log Pis Std 14.099 +trainer/Log Pis Max 73.375 +trainer/Log Pis Min -7.71667 +trainer/policy/mean Mean -0.0486505 +trainer/policy/mean Std 0.90802 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.89529 +trainer/policy/normal/std Std 0.658157 +trainer/policy/normal/std Max 5.57707 +trainer/policy/normal/std Min 0.277941 +trainer/policy/normal/log_std Mean 1.02439 +trainer/policy/normal/log_std Std 0.317411 +trainer/policy/normal/log_std Max 1.71866 +trainer/policy/normal/log_std Min -1.28035 +eval/num steps total 331752 +eval/num paths total 333 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123711 +eval/Actions Std 0.89084 +eval/Actions Max 0.999982 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51959 +time/logging (s) 0.00371455 +time/sampling batch (s) 0.276552 +time/saving (s) 0.00338563 +time/training (s) 6.9797 +time/epoch (s) 9.78294 +time/total (s) 3475.76 +Epoch -668 +---------------------------------- --------------- +2022-05-10 14:08:45.264822 PDT | [1] Epoch -667 finished +---------------------------------- --------------- +epoch -667 +replay_buffer/size 999033 +trainer/num train calls 334000 +trainer/Policy Loss -19.7608 +trainer/Log Pis Mean 24.6183 +trainer/Log Pis Std 13.511 +trainer/Log Pis Max 68.7414 +trainer/Log Pis Min -10.2385 +trainer/policy/mean Mean -0.0476549 +trainer/policy/mean Std 0.906664 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81519 +trainer/policy/normal/std Std 0.611121 +trainer/policy/normal/std Max 5.27959 +trainer/policy/normal/std Min 0.393964 +trainer/policy/normal/log_std Mean 1.00151 +trainer/policy/normal/log_std Std 0.28995 +trainer/policy/normal/log_std Max 1.66385 +trainer/policy/normal/log_std Min -0.931495 +eval/num steps total 332752 +eval/num paths total 334 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118154 +eval/Actions Std 0.901758 +eval/Actions Max 0.999995 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70502 +time/logging (s) 0.00420625 +time/sampling batch (s) 0.276668 +time/saving (s) 0.0040058 +time/training (s) 7.22613 +time/epoch (s) 10.216 +time/total (s) 3485.98 +Epoch -667 +---------------------------------- --------------- +2022-05-10 14:08:56.354954 PDT | [1] Epoch -666 finished +---------------------------------- --------------- +epoch -666 +replay_buffer/size 999033 +trainer/num train calls 335000 +trainer/Policy Loss -19.322 +trainer/Log Pis Mean 24.6662 +trainer/Log Pis Std 12.9485 +trainer/Log Pis Max 67.1048 +trainer/Log Pis Min -4.28196 +trainer/policy/mean Mean -0.0204637 +trainer/policy/mean Std 0.906624 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81813 +trainer/policy/normal/std Std 0.62617 +trainer/policy/normal/std Max 5.3881 +trainer/policy/normal/std Min 0.3206 +trainer/policy/normal/log_std Mean 1.00068 +trainer/policy/normal/log_std Std 0.299077 +trainer/policy/normal/log_std Max 1.68419 +trainer/policy/normal/log_std Min -1.13756 +eval/num steps total 333752 +eval/num paths total 335 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.399509 +eval/Actions Std 0.87222 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57557 +time/logging (s) 0.00372945 +time/sampling batch (s) 0.276966 +time/saving (s) 0.00353173 +time/training (s) 8.20814 +time/epoch (s) 11.0679 +time/total (s) 3497.05 +Epoch -666 +---------------------------------- --------------- +2022-05-10 14:09:06.743954 PDT | [1] Epoch -665 finished +---------------------------------- --------------- +epoch -665 +replay_buffer/size 999033 +trainer/num train calls 336000 +trainer/Policy Loss -19.5378 +trainer/Log Pis Mean 24.6051 +trainer/Log Pis Std 13.078 +trainer/Log Pis Max 75.7326 +trainer/Log Pis Min -8.27919 +trainer/policy/mean Mean -0.0343389 +trainer/policy/mean Std 0.90589 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.773 +trainer/policy/normal/std Std 0.607272 +trainer/policy/normal/std Max 5.25787 +trainer/policy/normal/std Min 0.28804 +trainer/policy/normal/log_std Mean 0.985639 +trainer/policy/normal/log_std Std 0.294009 +trainer/policy/normal/log_std Max 1.65973 +trainer/policy/normal/log_std Min -1.24466 +eval/num steps total 334752 +eval/num paths total 336 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.242834 +eval/Actions Std 0.919154 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73956 +time/logging (s) 0.00370401 +time/sampling batch (s) 0.276264 +time/saving (s) 0.00338677 +time/training (s) 7.34468 +time/epoch (s) 10.3676 +time/total (s) 3507.42 +Epoch -665 +---------------------------------- --------------- +2022-05-10 14:09:16.526570 PDT | [1] Epoch -664 finished +---------------------------------- --------------- +epoch -664 +replay_buffer/size 999033 +trainer/num train calls 337000 +trainer/Policy Loss -18.2073 +trainer/Log Pis Mean 22.9318 +trainer/Log Pis Std 12.7325 +trainer/Log Pis Max 70.9593 +trainer/Log Pis Min -11.6678 +trainer/policy/mean Mean -0.0315064 +trainer/policy/mean Std 0.901953 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79092 +trainer/policy/normal/std Std 0.629169 +trainer/policy/normal/std Max 6.03966 +trainer/policy/normal/std Min 0.276452 +trainer/policy/normal/log_std Mean 0.989342 +trainer/policy/normal/log_std Std 0.307406 +trainer/policy/normal/log_std Max 1.79835 +trainer/policy/normal/log_std Min -1.28572 +eval/num steps total 335752 +eval/num paths total 337 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118356 +eval/Actions Std 0.892459 +eval/Actions Max 0.999992 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59896 +time/logging (s) 0.00382717 +time/sampling batch (s) 0.277252 +time/saving (s) 0.00344108 +time/training (s) 6.87788 +time/epoch (s) 9.76136 +time/total (s) 3517.19 +Epoch -664 +---------------------------------- --------------- +2022-05-10 14:09:26.225328 PDT | [1] Epoch -663 finished +---------------------------------- --------------- +epoch -663 +replay_buffer/size 999033 +trainer/num train calls 338000 +trainer/Policy Loss -19.6001 +trainer/Log Pis Mean 24.0411 +trainer/Log Pis Std 13.5042 +trainer/Log Pis Max 79.1067 +trainer/Log Pis Min -7.66591 +trainer/policy/mean Mean -0.0454605 +trainer/policy/mean Std 0.90591 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82207 +trainer/policy/normal/std Std 0.636132 +trainer/policy/normal/std Max 5.79913 +trainer/policy/normal/std Min 0.357246 +trainer/policy/normal/log_std Mean 1.00077 +trainer/policy/normal/log_std Std 0.30617 +trainer/policy/normal/log_std Max 1.75771 +trainer/policy/normal/log_std Min -1.02933 +eval/num steps total 336752 +eval/num paths total 338 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0451597 +eval/Actions Std 0.880945 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49339 +time/logging (s) 0.00403209 +time/sampling batch (s) 0.285091 +time/saving (s) 0.00372597 +time/training (s) 6.89081 +time/epoch (s) 9.67705 +time/total (s) 3526.87 +Epoch -663 +---------------------------------- --------------- +2022-05-10 14:09:37.128036 PDT | [1] Epoch -662 finished +---------------------------------- --------------- +epoch -662 +replay_buffer/size 999033 +trainer/num train calls 339000 +trainer/Policy Loss -20.3316 +trainer/Log Pis Mean 24.2419 +trainer/Log Pis Std 12.6048 +trainer/Log Pis Max 71.2566 +trainer/Log Pis Min -9.01134 +trainer/policy/mean Mean -0.0359832 +trainer/policy/mean Std 0.908659 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85884 +trainer/policy/normal/std Std 0.644913 +trainer/policy/normal/std Max 5.95967 +trainer/policy/normal/std Min 0.19209 +trainer/policy/normal/log_std Mean 1.01354 +trainer/policy/normal/log_std Std 0.307406 +trainer/policy/normal/log_std Max 1.78502 +trainer/policy/normal/log_std Min -1.64979 +eval/num steps total 337752 +eval/num paths total 339 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.175159 +eval/Actions Std 0.803225 +eval/Actions Max 0.999972 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75095 +time/logging (s) 0.00410966 +time/sampling batch (s) 0.285684 +time/saving (s) 0.00382658 +time/training (s) 7.836 +time/epoch (s) 10.8806 +time/total (s) 3537.75 +Epoch -662 +---------------------------------- --------------- +2022-05-10 14:09:47.048429 PDT | [1] Epoch -661 finished +---------------------------------- --------------- +epoch -661 +replay_buffer/size 999033 +trainer/num train calls 340000 +trainer/Policy Loss -20.4623 +trainer/Log Pis Mean 23.6625 +trainer/Log Pis Std 13.147 +trainer/Log Pis Max 57.4382 +trainer/Log Pis Min -7.82749 +trainer/policy/mean Mean -0.0506915 +trainer/policy/mean Std 0.902396 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.7869 +trainer/policy/normal/std Std 0.609494 +trainer/policy/normal/std Max 5.38811 +trainer/policy/normal/std Min 0.364403 +trainer/policy/normal/log_std Mean 0.990713 +trainer/policy/normal/log_std Std 0.293786 +trainer/policy/normal/log_std Max 1.68419 +trainer/policy/normal/log_std Min -1.0095 +eval/num steps total 338293 +eval/num paths total 340 +eval/path length Mean 541 +eval/path length Std 0 +eval/path length Max 541 +eval/path length Min 541 +eval/Rewards Mean 0.00184843 +eval/Rewards Std 0.0429536 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0245699 +eval/Actions Std 0.900501 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.31571 +time/logging (s) 0.00257362 +time/sampling batch (s) 0.27758 +time/saving (s) 0.00372913 +time/training (s) 7.29732 +time/epoch (s) 9.89691 +time/total (s) 3547.65 +Epoch -661 +---------------------------------- --------------- +2022-05-10 14:09:57.079753 PDT | [1] Epoch -660 finished +---------------------------------- --------------- +epoch -660 +replay_buffer/size 999033 +trainer/num train calls 341000 +trainer/Policy Loss -19.478 +trainer/Log Pis Mean 24.6496 +trainer/Log Pis Std 13.1384 +trainer/Log Pis Max 70.7232 +trainer/Log Pis Min -8.61381 +trainer/policy/mean Mean -0.0210058 +trainer/policy/mean Std 0.906681 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85878 +trainer/policy/normal/std Std 0.636154 +trainer/policy/normal/std Max 5.80755 +trainer/policy/normal/std Min 0.402329 +trainer/policy/normal/log_std Mean 1.01619 +trainer/policy/normal/log_std Std 0.290749 +trainer/policy/normal/log_std Max 1.75916 +trainer/policy/normal/log_std Min -0.910484 +eval/num steps total 339293 +eval/num paths total 341 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0149887 +eval/Actions Std 0.910612 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76762 +time/logging (s) 0.00371319 +time/sampling batch (s) 0.528256 +time/saving (s) 0.00346457 +time/training (s) 6.70805 +time/epoch (s) 10.0111 +time/total (s) 3557.66 +Epoch -660 +---------------------------------- --------------- +2022-05-10 14:10:06.062660 PDT | [1] Epoch -659 finished +---------------------------------- --------------- +epoch -659 +replay_buffer/size 999033 +trainer/num train calls 342000 +trainer/Policy Loss -19.6956 +trainer/Log Pis Mean 24.2262 +trainer/Log Pis Std 12.7873 +trainer/Log Pis Max 71.7278 +trainer/Log Pis Min -10.3351 +trainer/policy/mean Mean -0.025767 +trainer/policy/mean Std 0.908652 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82087 +trainer/policy/normal/std Std 0.616012 +trainer/policy/normal/std Max 6.1446 +trainer/policy/normal/std Min 0.354644 +trainer/policy/normal/log_std Mean 1.00194 +trainer/policy/normal/log_std Std 0.300234 +trainer/policy/normal/log_std Max 1.81557 +trainer/policy/normal/log_std Min -1.03664 +eval/num steps total 340293 +eval/num paths total 342 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.185669 +eval/Actions Std 0.894565 +eval/Actions Max 0.99999 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.632 +time/logging (s) 0.00370655 +time/sampling batch (s) 0.276406 +time/saving (s) 0.0034511 +time/training (s) 6.04597 +time/epoch (s) 8.96154 +time/total (s) 3566.63 +Epoch -659 +---------------------------------- --------------- +2022-05-10 14:10:15.644276 PDT | [1] Epoch -658 finished +---------------------------------- --------------- +epoch -658 +replay_buffer/size 999033 +trainer/num train calls 343000 +trainer/Policy Loss -20.5377 +trainer/Log Pis Mean 24.3927 +trainer/Log Pis Std 13.3646 +trainer/Log Pis Max 83.8667 +trainer/Log Pis Min -6.75608 +trainer/policy/mean Mean -0.0556709 +trainer/policy/mean Std 0.900993 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86269 +trainer/policy/normal/std Std 0.662131 +trainer/policy/normal/std Max 5.90691 +trainer/policy/normal/std Min 0.339349 +trainer/policy/normal/log_std Mean 1.01223 +trainer/policy/normal/log_std Std 0.319394 +trainer/policy/normal/log_std Max 1.77612 +trainer/policy/normal/log_std Min -1.08073 +eval/num steps total 341293 +eval/num paths total 343 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.265767 +eval/Actions Std 0.873993 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54722 +time/logging (s) 0.00370496 +time/sampling batch (s) 0.276662 +time/saving (s) 0.0034047 +time/training (s) 6.72911 +time/epoch (s) 9.5601 +time/total (s) 3576.19 +Epoch -658 +---------------------------------- --------------- +2022-05-10 14:10:25.955715 PDT | [1] Epoch -657 finished +---------------------------------- --------------- +epoch -657 +replay_buffer/size 999033 +trainer/num train calls 344000 +trainer/Policy Loss -19.9217 +trainer/Log Pis Mean 23.969 +trainer/Log Pis Std 13.1822 +trainer/Log Pis Max 65.9315 +trainer/Log Pis Min -7.64159 +trainer/policy/mean Mean -0.0390188 +trainer/policy/mean Std 0.906436 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80775 +trainer/policy/normal/std Std 0.619967 +trainer/policy/normal/std Max 4.9841 +trainer/policy/normal/std Min 0.251539 +trainer/policy/normal/log_std Mean 0.996257 +trainer/policy/normal/log_std Std 0.306928 +trainer/policy/normal/log_std Max 1.60625 +trainer/policy/normal/log_std Min -1.38016 +eval/num steps total 342293 +eval/num paths total 344 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0175041 +eval/Actions Std 0.909391 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6553 +time/logging (s) 0.00373979 +time/sampling batch (s) 0.275839 +time/saving (s) 0.00347618 +time/training (s) 7.35176 +time/epoch (s) 10.2901 +time/total (s) 3586.48 +Epoch -657 +---------------------------------- --------------- +2022-05-10 14:10:35.215728 PDT | [1] Epoch -656 finished +---------------------------------- --------------- +epoch -656 +replay_buffer/size 999033 +trainer/num train calls 345000 +trainer/Policy Loss -18.9797 +trainer/Log Pis Mean 25.0163 +trainer/Log Pis Std 13.0581 +trainer/Log Pis Max 74.3909 +trainer/Log Pis Min -8.10852 +trainer/policy/mean Mean -0.0447547 +trainer/policy/mean Std 0.910427 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82439 +trainer/policy/normal/std Std 0.621001 +trainer/policy/normal/std Max 5.15254 +trainer/policy/normal/std Min 0.293764 +trainer/policy/normal/log_std Mean 1.00236 +trainer/policy/normal/log_std Std 0.304131 +trainer/policy/normal/log_std Max 1.63949 +trainer/policy/normal/log_std Min -1.22498 +eval/num steps total 343293 +eval/num paths total 345 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.309947 +eval/Actions Std 0.877131 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39648 +time/logging (s) 0.00403568 +time/sampling batch (s) 0.274189 +time/saving (s) 0.00394756 +time/training (s) 6.56027 +time/epoch (s) 9.23893 +time/total (s) 3595.73 +Epoch -656 +---------------------------------- --------------- +2022-05-10 14:10:45.674205 PDT | [1] Epoch -655 finished +---------------------------------- --------------- +epoch -655 +replay_buffer/size 999033 +trainer/num train calls 346000 +trainer/Policy Loss -19.6608 +trainer/Log Pis Mean 23.5109 +trainer/Log Pis Std 13.2328 +trainer/Log Pis Max 65.7123 +trainer/Log Pis Min -13.1026 +trainer/policy/mean Mean -0.0158712 +trainer/policy/mean Std 0.903235 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80008 +trainer/policy/normal/std Std 0.629958 +trainer/policy/normal/std Max 6.0396 +trainer/policy/normal/std Min 0.344744 +trainer/policy/normal/log_std Mean 0.99274 +trainer/policy/normal/log_std Std 0.307954 +trainer/policy/normal/log_std Max 1.79834 +trainer/policy/normal/log_std Min -1.06495 +eval/num steps total 344293 +eval/num paths total 346 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0340505 +eval/Actions Std 0.909232 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5559 +time/logging (s) 0.00375702 +time/sampling batch (s) 0.525407 +time/saving (s) 0.0033564 +time/training (s) 7.3481 +time/epoch (s) 10.4365 +time/total (s) 3606.17 +Epoch -655 +---------------------------------- --------------- +2022-05-10 14:10:56.072663 PDT | [1] Epoch -654 finished +---------------------------------- --------------- +epoch -654 +replay_buffer/size 999033 +trainer/num train calls 347000 +trainer/Policy Loss -19.7198 +trainer/Log Pis Mean 25.5698 +trainer/Log Pis Std 13.5386 +trainer/Log Pis Max 75.0841 +trainer/Log Pis Min -7.78239 +trainer/policy/mean Mean -0.0497604 +trainer/policy/mean Std 0.912214 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84691 +trainer/policy/normal/std Std 0.61028 +trainer/policy/normal/std Max 5.36416 +trainer/policy/normal/std Min 0.265402 +trainer/policy/normal/log_std Mean 1.01306 +trainer/policy/normal/log_std Std 0.290616 +trainer/policy/normal/log_std Max 1.67974 +trainer/policy/normal/log_std Min -1.32651 +eval/num steps total 345293 +eval/num paths total 347 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.39051 +eval/Actions Std 0.845312 +eval/Actions Max 0.999971 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39611 +time/logging (s) 0.00375945 +time/sampling batch (s) 0.52499 +time/saving (s) 0.00349945 +time/training (s) 7.4488 +time/epoch (s) 10.3772 +time/total (s) 3616.55 +Epoch -654 +---------------------------------- --------------- +2022-05-10 14:11:06.123079 PDT | [1] Epoch -653 finished +---------------------------------- --------------- +epoch -653 +replay_buffer/size 999033 +trainer/num train calls 348000 +trainer/Policy Loss -19.6204 +trainer/Log Pis Mean 23.9055 +trainer/Log Pis Std 13.3031 +trainer/Log Pis Max 71.016 +trainer/Log Pis Min -7.42286 +trainer/policy/mean Mean -0.0426228 +trainer/policy/mean Std 0.906913 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83571 +trainer/policy/normal/std Std 0.602525 +trainer/policy/normal/std Max 5.27791 +trainer/policy/normal/std Min 0.396877 +trainer/policy/normal/log_std Mean 1.01148 +trainer/policy/normal/log_std Std 0.27575 +trainer/policy/normal/log_std Max 1.66353 +trainer/policy/normal/log_std Min -0.924128 +eval/num steps total 346293 +eval/num paths total 348 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.122661 +eval/Actions Std 0.91509 +eval/Actions Max 0.999982 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69026 +time/logging (s) 0.00386106 +time/sampling batch (s) 0.278245 +time/saving (s) 0.00342482 +time/training (s) 7.05324 +time/epoch (s) 10.029 +time/total (s) 3626.58 +Epoch -653 +---------------------------------- --------------- +2022-05-10 14:11:17.429666 PDT | [1] Epoch -652 finished +---------------------------------- --------------- +epoch -652 +replay_buffer/size 999033 +trainer/num train calls 349000 +trainer/Policy Loss -20.1288 +trainer/Log Pis Mean 23.7673 +trainer/Log Pis Std 13.3462 +trainer/Log Pis Max 74.4289 +trainer/Log Pis Min -11.1319 +trainer/policy/mean Mean -0.0233649 +trainer/policy/mean Std 0.902077 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79271 +trainer/policy/normal/std Std 0.633968 +trainer/policy/normal/std Max 5.94914 +trainer/policy/normal/std Min 0.325127 +trainer/policy/normal/log_std Mean 0.990702 +trainer/policy/normal/log_std Std 0.30216 +trainer/policy/normal/log_std Max 1.78325 +trainer/policy/normal/log_std Min -1.12354 +eval/num steps total 347293 +eval/num paths total 349 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.472509 +eval/Actions Std 0.864028 +eval/Actions Max 0.999994 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78176 +time/logging (s) 0.00384906 +time/sampling batch (s) 0.278865 +time/saving (s) 0.00344214 +time/training (s) 8.21712 +time/epoch (s) 11.285 +time/total (s) 3637.87 +Epoch -652 +---------------------------------- --------------- +2022-05-10 14:11:28.268185 PDT | [1] Epoch -651 finished +---------------------------------- --------------- +epoch -651 +replay_buffer/size 999033 +trainer/num train calls 350000 +trainer/Policy Loss -19.7166 +trainer/Log Pis Mean 24.9583 +trainer/Log Pis Std 13.4932 +trainer/Log Pis Max 62.4401 +trainer/Log Pis Min -9.29986 +trainer/policy/mean Mean -0.0287147 +trainer/policy/mean Std 0.90615 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78757 +trainer/policy/normal/std Std 0.628907 +trainer/policy/normal/std Max 6.62628 +trainer/policy/normal/std Min 0.348266 +trainer/policy/normal/log_std Mean 0.98864 +trainer/policy/normal/log_std Std 0.30467 +trainer/policy/normal/log_std Max 1.89104 +trainer/policy/normal/log_std Min -1.05479 +eval/num steps total 348293 +eval/num paths total 350 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110175 +eval/Actions Std 0.895757 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56046 +time/logging (s) 0.00414034 +time/sampling batch (s) 0.780239 +time/saving (s) 0.00411779 +time/training (s) 7.46818 +time/epoch (s) 10.8171 +time/total (s) 3648.69 +Epoch -651 +---------------------------------- --------------- +2022-05-10 14:11:38.073356 PDT | [1] Epoch -650 finished +---------------------------------- --------------- +epoch -650 +replay_buffer/size 999033 +trainer/num train calls 351000 +trainer/Policy Loss -19.7157 +trainer/Log Pis Mean 24.7233 +trainer/Log Pis Std 13.164 +trainer/Log Pis Max 70.4894 +trainer/Log Pis Min -5.48104 +trainer/policy/mean Mean -0.0344559 +trainer/policy/mean Std 0.904723 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81748 +trainer/policy/normal/std Std 0.62554 +trainer/policy/normal/std Max 5.20946 +trainer/policy/normal/std Min 0.323623 +trainer/policy/normal/log_std Mean 1.00058 +trainer/policy/normal/log_std Std 0.298305 +trainer/policy/normal/log_std Max 1.65048 +trainer/policy/normal/log_std Min -1.12818 +eval/num steps total 349293 +eval/num paths total 351 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.316421 +eval/Actions Std 0.90236 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7344 +time/logging (s) 0.00377294 +time/sampling batch (s) 0.530563 +time/saving (s) 0.00364201 +time/training (s) 6.51046 +time/epoch (s) 9.78284 +time/total (s) 3658.47 +Epoch -650 +---------------------------------- --------------- +2022-05-10 14:11:49.055402 PDT | [1] Epoch -649 finished +---------------------------------- --------------- +epoch -649 +replay_buffer/size 999033 +trainer/num train calls 352000 +trainer/Policy Loss -19.6574 +trainer/Log Pis Mean 24.938 +trainer/Log Pis Std 13.4162 +trainer/Log Pis Max 66.3515 +trainer/Log Pis Min -10.5169 +trainer/policy/mean Mean -0.0346563 +trainer/policy/mean Std 0.906337 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79418 +trainer/policy/normal/std Std 0.605008 +trainer/policy/normal/std Max 5.50561 +trainer/policy/normal/std Min 0.365408 +trainer/policy/normal/log_std Mean 0.994962 +trainer/policy/normal/log_std Std 0.284507 +trainer/policy/normal/log_std Max 1.70577 +trainer/policy/normal/log_std Min -1.00674 +eval/num steps total 350293 +eval/num paths total 352 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127213 +eval/Actions Std 0.903498 +eval/Actions Max 0.999995 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73004 +time/logging (s) 0.00395233 +time/sampling batch (s) 0.28047 +time/saving (s) 0.00360228 +time/training (s) 7.94212 +time/epoch (s) 10.9602 +time/total (s) 3669.44 +Epoch -649 +---------------------------------- --------------- +2022-05-10 14:11:59.150709 PDT | [1] Epoch -648 finished +---------------------------------- --------------- +epoch -648 +replay_buffer/size 999033 +trainer/num train calls 353000 +trainer/Policy Loss -19.7067 +trainer/Log Pis Mean 24.7441 +trainer/Log Pis Std 13.0323 +trainer/Log Pis Max 63.2341 +trainer/Log Pis Min -8.87461 +trainer/policy/mean Mean -0.0145883 +trainer/policy/mean Std 0.908802 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81963 +trainer/policy/normal/std Std 0.639948 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.338324 +trainer/policy/normal/log_std Mean 1.00005 +trainer/policy/normal/log_std Std 0.3048 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.08375 +eval/num steps total 351293 +eval/num paths total 353 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0430907 +eval/Actions Std 0.912785 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7775 +time/logging (s) 0.00381983 +time/sampling batch (s) 0.282563 +time/saving (s) 0.00349164 +time/training (s) 7.00583 +time/epoch (s) 10.0732 +time/total (s) 3679.51 +Epoch -648 +---------------------------------- --------------- +2022-05-10 14:12:08.271813 PDT | [1] Epoch -647 finished +---------------------------------- --------------- +epoch -647 +replay_buffer/size 999033 +trainer/num train calls 354000 +trainer/Policy Loss -20.3038 +trainer/Log Pis Mean 24.6704 +trainer/Log Pis Std 13.1273 +trainer/Log Pis Max 67.9189 +trainer/Log Pis Min -7.10082 +trainer/policy/mean Mean -0.0462407 +trainer/policy/mean Std 0.906059 +trainer/policy/mean Max 0.99997 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.89502 +trainer/policy/normal/std Std 0.646563 +trainer/policy/normal/std Max 5.51292 +trainer/policy/normal/std Min 0.257176 +trainer/policy/normal/log_std Mean 1.0274 +trainer/policy/normal/log_std Std 0.300158 +trainer/policy/normal/log_std Max 1.70709 +trainer/policy/normal/log_std Min -1.358 +eval/num steps total 352293 +eval/num paths total 354 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.256546 +eval/Actions Std 0.912859 +eval/Actions Max 0.999931 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61122 +time/logging (s) 0.00370114 +time/sampling batch (s) 0.279093 +time/saving (s) 0.00342421 +time/training (s) 6.20176 +time/epoch (s) 9.09919 +time/total (s) 3688.61 +Epoch -647 +---------------------------------- --------------- +2022-05-10 14:12:18.122593 PDT | [1] Epoch -646 finished +---------------------------------- --------------- +epoch -646 +replay_buffer/size 999033 +trainer/num train calls 355000 +trainer/Policy Loss -18.8048 +trainer/Log Pis Mean 23.9163 +trainer/Log Pis Std 12.5029 +trainer/Log Pis Max 77.0881 +trainer/Log Pis Min -5.16382 +trainer/policy/mean Mean -0.02587 +trainer/policy/mean Std 0.902827 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85096 +trainer/policy/normal/std Std 0.615197 +trainer/policy/normal/std Max 5.99022 +trainer/policy/normal/std Min 0.385786 +trainer/policy/normal/log_std Mean 1.01514 +trainer/policy/normal/log_std Std 0.28454 +trainer/policy/normal/log_std Max 1.79013 +trainer/policy/normal/log_std Min -0.952473 +eval/num steps total 353293 +eval/num paths total 355 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102608 +eval/Actions Std 0.899651 +eval/Actions Max 0.999994 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49855 +time/logging (s) 0.00369568 +time/sampling batch (s) 0.279776 +time/saving (s) 0.00347658 +time/training (s) 7.04355 +time/epoch (s) 9.82904 +time/total (s) 3698.45 +Epoch -646 +---------------------------------- --------------- +2022-05-10 14:12:27.971661 PDT | [1] Epoch -645 finished +---------------------------------- --------------- +epoch -645 +replay_buffer/size 999033 +trainer/num train calls 356000 +trainer/Policy Loss -20.1642 +trainer/Log Pis Mean 24.5369 +trainer/Log Pis Std 13.6008 +trainer/Log Pis Max 73.052 +trainer/Log Pis Min -6.84367 +trainer/policy/mean Mean -0.0237396 +trainer/policy/mean Std 0.906817 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.81711 +trainer/policy/normal/std Std 0.639199 +trainer/policy/normal/std Max 5.68814 +trainer/policy/normal/std Min 0.272464 +trainer/policy/normal/log_std Mean 0.998975 +trainer/policy/normal/log_std Std 0.30471 +trainer/policy/normal/log_std Max 1.73838 +trainer/policy/normal/log_std Min -1.30025 +eval/num steps total 354293 +eval/num paths total 356 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0492531 +eval/Actions Std 0.93035 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70044 +time/logging (s) 0.00399185 +time/sampling batch (s) 0.279236 +time/saving (s) 0.00396114 +time/training (s) 6.84013 +time/epoch (s) 9.82777 +time/total (s) 3708.28 +Epoch -645 +---------------------------------- --------------- +2022-05-10 14:12:38.924346 PDT | [1] Epoch -644 finished +---------------------------------- --------------- +epoch -644 +replay_buffer/size 999033 +trainer/num train calls 357000 +trainer/Policy Loss -20.7567 +trainer/Log Pis Mean 25.4735 +trainer/Log Pis Std 13.4868 +trainer/Log Pis Max 69.7037 +trainer/Log Pis Min -3.47934 +trainer/policy/mean Mean -0.0423517 +trainer/policy/mean Std 0.911499 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84469 +trainer/policy/normal/std Std 0.601346 +trainer/policy/normal/std Max 6.00829 +trainer/policy/normal/std Min 0.325214 +trainer/policy/normal/log_std Mean 1.01324 +trainer/policy/normal/log_std Std 0.286779 +trainer/policy/normal/log_std Max 1.79314 +trainer/policy/normal/log_std Min -1.12327 +eval/num steps total 355293 +eval/num paths total 357 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.137617 +eval/Actions Std 0.899582 +eval/Actions Max 0.999947 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45616 +time/logging (s) 0.00366789 +time/sampling batch (s) 0.278603 +time/saving (s) 0.0034061 +time/training (s) 8.18862 +time/epoch (s) 10.9305 +time/total (s) 3719.21 +Epoch -644 +---------------------------------- --------------- +2022-05-10 14:12:49.139890 PDT | [1] Epoch -643 finished +---------------------------------- --------------- +epoch -643 +replay_buffer/size 999033 +trainer/num train calls 358000 +trainer/Policy Loss -19.1962 +trainer/Log Pis Mean 24.8443 +trainer/Log Pis Std 12.7453 +trainer/Log Pis Max 66.1646 +trainer/Log Pis Min -5.94585 +trainer/policy/mean Mean -0.0283318 +trainer/policy/mean Std 0.907733 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85641 +trainer/policy/normal/std Std 0.616488 +trainer/policy/normal/std Max 5.76287 +trainer/policy/normal/std Min 0.361539 +trainer/policy/normal/log_std Mean 1.01552 +trainer/policy/normal/log_std Std 0.295789 +trainer/policy/normal/log_std Max 1.75143 +trainer/policy/normal/log_std Min -1.01739 +eval/num steps total 356293 +eval/num paths total 358 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0651652 +eval/Actions Std 0.880766 +eval/Actions Max 0.999972 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82646 +time/logging (s) 0.00367845 +time/sampling batch (s) 0.278371 +time/saving (s) 0.00337986 +time/training (s) 7.08217 +time/epoch (s) 10.1941 +time/total (s) 3729.41 +Epoch -643 +---------------------------------- --------------- +2022-05-10 14:12:59.189261 PDT | [1] Epoch -642 finished +---------------------------------- --------------- +epoch -642 +replay_buffer/size 999033 +trainer/num train calls 359000 +trainer/Policy Loss -19.3964 +trainer/Log Pis Mean 23.9118 +trainer/Log Pis Std 13.9918 +trainer/Log Pis Max 68.5789 +trainer/Log Pis Min -8.63223 +trainer/policy/mean Mean -0.0496738 +trainer/policy/mean Std 0.901957 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81659 +trainer/policy/normal/std Std 0.629173 +trainer/policy/normal/std Max 5.09409 +trainer/policy/normal/std Min 0.36747 +trainer/policy/normal/log_std Mean 0.999452 +trainer/policy/normal/log_std Std 0.302713 +trainer/policy/normal/log_std Max 1.62808 +trainer/policy/normal/log_std Min -1.00111 +eval/num steps total 357293 +eval/num paths total 359 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0760242 +eval/Actions Std 0.900327 +eval/Actions Max 0.999992 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57067 +time/logging (s) 0.00377923 +time/sampling batch (s) 0.775506 +time/saving (s) 0.00341772 +time/training (s) 6.6749 +time/epoch (s) 10.0283 +time/total (s) 3739.44 +Epoch -642 +---------------------------------- --------------- +2022-05-10 14:13:10.805626 PDT | [1] Epoch -641 finished +---------------------------------- --------------- +epoch -641 +replay_buffer/size 999033 +trainer/num train calls 360000 +trainer/Policy Loss -20.4593 +trainer/Log Pis Mean 24.2195 +trainer/Log Pis Std 13.1772 +trainer/Log Pis Max 68.7358 +trainer/Log Pis Min -8.26075 +trainer/policy/mean Mean -0.0318037 +trainer/policy/mean Std 0.906763 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80614 +trainer/policy/normal/std Std 0.623007 +trainer/policy/normal/std Max 6.37564 +trainer/policy/normal/std Min 0.352735 +trainer/policy/normal/log_std Mean 0.996704 +trainer/policy/normal/log_std Std 0.297321 +trainer/policy/normal/log_std Max 1.85248 +trainer/policy/normal/log_std Min -1.04204 +eval/num steps total 358293 +eval/num paths total 360 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0697587 +eval/Actions Std 0.875763 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48368 +time/logging (s) 0.00370478 +time/sampling batch (s) 0.525759 +time/saving (s) 0.00338258 +time/training (s) 8.57848 +time/epoch (s) 11.595 +time/total (s) 3751.04 +Epoch -641 +---------------------------------- --------------- +2022-05-10 14:13:21.925264 PDT | [1] Epoch -640 finished +---------------------------------- --------------- +epoch -640 +replay_buffer/size 999033 +trainer/num train calls 361000 +trainer/Policy Loss -20.4012 +trainer/Log Pis Mean 25.0782 +trainer/Log Pis Std 13.5767 +trainer/Log Pis Max 68.6272 +trainer/Log Pis Min -7.61807 +trainer/policy/mean Mean -0.0365889 +trainer/policy/mean Std 0.90984 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82408 +trainer/policy/normal/std Std 0.632359 +trainer/policy/normal/std Max 5.20723 +trainer/policy/normal/std Min 0.346216 +trainer/policy/normal/log_std Mean 1.00121 +trainer/policy/normal/log_std Std 0.308258 +trainer/policy/normal/log_std Max 1.65005 +trainer/policy/normal/log_std Min -1.06069 +eval/num steps total 359293 +eval/num paths total 361 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126557 +eval/Actions Std 0.872167 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66805 +time/logging (s) 0.0041286 +time/sampling batch (s) 0.528145 +time/saving (s) 0.00403464 +time/training (s) 7.89407 +time/epoch (s) 11.0984 +time/total (s) 3762.14 +Epoch -640 +---------------------------------- --------------- +2022-05-10 14:13:32.426394 PDT | [1] Epoch -639 finished +---------------------------------- --------------- +epoch -639 +replay_buffer/size 999033 +trainer/num train calls 362000 +trainer/Policy Loss -19.2339 +trainer/Log Pis Mean 25.6761 +trainer/Log Pis Std 13.8763 +trainer/Log Pis Max 73.1282 +trainer/Log Pis Min -8.24317 +trainer/policy/mean Mean -0.0336108 +trainer/policy/mean Std 0.910484 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87558 +trainer/policy/normal/std Std 0.616245 +trainer/policy/normal/std Max 6.08448 +trainer/policy/normal/std Min 0.285848 +trainer/policy/normal/log_std Mean 1.02331 +trainer/policy/normal/log_std Std 0.288765 +trainer/policy/normal/log_std Max 1.80574 +trainer/policy/normal/log_std Min -1.2523 +eval/num steps total 360293 +eval/num paths total 362 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0713892 +eval/Actions Std 0.921498 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64206 +time/logging (s) 0.00378754 +time/sampling batch (s) 0.275328 +time/saving (s) 0.0036768 +time/training (s) 7.55428 +time/epoch (s) 10.4791 +time/total (s) 3772.62 +Epoch -639 +---------------------------------- --------------- +2022-05-10 14:13:43.041513 PDT | [1] Epoch -638 finished +---------------------------------- --------------- +epoch -638 +replay_buffer/size 999033 +trainer/num train calls 363000 +trainer/Policy Loss -20.4472 +trainer/Log Pis Mean 25.0498 +trainer/Log Pis Std 13.1572 +trainer/Log Pis Max 74.6748 +trainer/Log Pis Min -10.8362 +trainer/policy/mean Mean -0.0322763 +trainer/policy/mean Std 0.907744 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.90363 +trainer/policy/normal/std Std 0.647385 +trainer/policy/normal/std Max 5.76021 +trainer/policy/normal/std Min 0.347986 +trainer/policy/normal/log_std Mean 1.02981 +trainer/policy/normal/log_std Std 0.303621 +trainer/policy/normal/log_std Max 1.75097 +trainer/policy/normal/log_std Min -1.05559 +eval/num steps total 361293 +eval/num paths total 363 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.185068 +eval/Actions Std 0.879166 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63664 +time/logging (s) 0.00374609 +time/sampling batch (s) 0.275043 +time/saving (s) 0.00341725 +time/training (s) 7.67495 +time/epoch (s) 10.5938 +time/total (s) 3783.22 +Epoch -638 +---------------------------------- --------------- +2022-05-10 14:13:53.429866 PDT | [1] Epoch -637 finished +---------------------------------- --------------- +epoch -637 +replay_buffer/size 999033 +trainer/num train calls 364000 +trainer/Policy Loss -18.8617 +trainer/Log Pis Mean 24.6194 +trainer/Log Pis Std 13.7973 +trainer/Log Pis Max 76.7706 +trainer/Log Pis Min -7.6622 +trainer/policy/mean Mean -0.0385879 +trainer/policy/mean Std 0.903758 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85984 +trainer/policy/normal/std Std 0.625973 +trainer/policy/normal/std Max 5.67579 +trainer/policy/normal/std Min 0.339441 +trainer/policy/normal/log_std Mean 1.01633 +trainer/policy/normal/log_std Std 0.2962 +trainer/policy/normal/log_std Max 1.73621 +trainer/policy/normal/log_std Min -1.08046 +eval/num steps total 362293 +eval/num paths total 364 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0331712 +eval/Actions Std 0.965137 +eval/Actions Max 0.999987 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65192 +time/logging (s) 0.00366214 +time/sampling batch (s) 0.27446 +time/saving (s) 0.00341537 +time/training (s) 7.43353 +time/epoch (s) 10.367 +time/total (s) 3793.59 +Epoch -637 +---------------------------------- --------------- +2022-05-10 14:14:03.729978 PDT | [1] Epoch -636 finished +---------------------------------- --------------- +epoch -636 +replay_buffer/size 999033 +trainer/num train calls 365000 +trainer/Policy Loss -19.9234 +trainer/Log Pis Mean 24.2848 +trainer/Log Pis Std 12.6511 +trainer/Log Pis Max 61.6317 +trainer/Log Pis Min -5.2987 +trainer/policy/mean Mean -0.0553011 +trainer/policy/mean Std 0.907914 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87752 +trainer/policy/normal/std Std 0.636249 +trainer/policy/normal/std Max 6.69027 +trainer/policy/normal/std Min 0.416922 +trainer/policy/normal/log_std Mean 1.02367 +trainer/policy/normal/log_std Std 0.285746 +trainer/policy/normal/log_std Max 1.90065 +trainer/policy/normal/log_std Min -0.874857 +eval/num steps total 363293 +eval/num paths total 365 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0625981 +eval/Actions Std 0.910538 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78101 +time/logging (s) 0.00382598 +time/sampling batch (s) 0.274067 +time/saving (s) 0.00344888 +time/training (s) 7.21668 +time/epoch (s) 10.279 +time/total (s) 3803.87 +Epoch -636 +---------------------------------- --------------- +2022-05-10 14:14:15.036401 PDT | [1] Epoch -635 finished +---------------------------------- --------------- +epoch -635 +replay_buffer/size 999033 +trainer/num train calls 366000 +trainer/Policy Loss -19.7506 +trainer/Log Pis Mean 24.2338 +trainer/Log Pis Std 13.0505 +trainer/Log Pis Max 73.4406 +trainer/Log Pis Min -5.22609 +trainer/policy/mean Mean -0.0622468 +trainer/policy/mean Std 0.901177 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81105 +trainer/policy/normal/std Std 0.624859 +trainer/policy/normal/std Max 4.956 +trainer/policy/normal/std Min 0.384905 +trainer/policy/normal/log_std Mean 0.99843 +trainer/policy/normal/log_std Std 0.296902 +trainer/policy/normal/log_std Max 1.6006 +trainer/policy/normal/log_std Min -0.954758 +eval/num steps total 364293 +eval/num paths total 366 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115221 +eval/Actions Std 0.902326 +eval/Actions Max 0.999975 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48743 +time/logging (s) 0.00429791 +time/sampling batch (s) 0.821624 +time/saving (s) 0.00438595 +time/training (s) 7.96653 +time/epoch (s) 11.2843 +time/total (s) 3815.16 +Epoch -635 +---------------------------------- --------------- +2022-05-10 14:14:24.540782 PDT | [1] Epoch -634 finished +---------------------------------- --------------- +epoch -634 +replay_buffer/size 999033 +trainer/num train calls 367000 +trainer/Policy Loss -20.3056 +trainer/Log Pis Mean 25.0603 +trainer/Log Pis Std 13.0893 +trainer/Log Pis Max 66.9648 +trainer/Log Pis Min -5.44726 +trainer/policy/mean Mean -0.0415891 +trainer/policy/mean Std 0.909301 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82306 +trainer/policy/normal/std Std 0.614807 +trainer/policy/normal/std Max 7.38358 +trainer/policy/normal/std Min 0.307353 +trainer/policy/normal/log_std Mean 1.00464 +trainer/policy/normal/log_std Std 0.288031 +trainer/policy/normal/log_std Max 1.99926 +trainer/policy/normal/log_std Min -1.17976 +eval/num steps total 365293 +eval/num paths total 367 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.42494 +eval/Actions Std 0.882932 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45154 +time/logging (s) 0.00420435 +time/sampling batch (s) 0.281137 +time/saving (s) 0.00392896 +time/training (s) 6.74099 +time/epoch (s) 9.4818 +time/total (s) 3824.64 +Epoch -634 +---------------------------------- --------------- +2022-05-10 14:14:34.476291 PDT | [1] Epoch -633 finished +---------------------------------- --------------- +epoch -633 +replay_buffer/size 999033 +trainer/num train calls 368000 +trainer/Policy Loss -19.5229 +trainer/Log Pis Mean 24.1865 +trainer/Log Pis Std 12.4155 +trainer/Log Pis Max 66.951 +trainer/Log Pis Min -7.14784 +trainer/policy/mean Mean -0.0441914 +trainer/policy/mean Std 0.906061 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80142 +trainer/policy/normal/std Std 0.613042 +trainer/policy/normal/std Max 5.20886 +trainer/policy/normal/std Min 0.311321 +trainer/policy/normal/log_std Mean 0.995619 +trainer/policy/normal/log_std Std 0.296052 +trainer/policy/normal/log_std Max 1.65036 +trainer/policy/normal/log_std Min -1.16693 +eval/num steps total 366293 +eval/num paths total 368 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.172249 +eval/Actions Std 0.844607 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75507 +time/logging (s) 0.00406474 +time/sampling batch (s) 0.277804 +time/saving (s) 0.00392944 +time/training (s) 6.87232 +time/epoch (s) 9.9132 +time/total (s) 3834.56 +Epoch -633 +---------------------------------- --------------- +2022-05-10 14:14:45.217228 PDT | [1] Epoch -632 finished +---------------------------------- --------------- +epoch -632 +replay_buffer/size 999033 +trainer/num train calls 369000 +trainer/Policy Loss -19.3038 +trainer/Log Pis Mean 24.0859 +trainer/Log Pis Std 13.5929 +trainer/Log Pis Max 62.736 +trainer/Log Pis Min -7.71183 +trainer/policy/mean Mean -0.0372541 +trainer/policy/mean Std 0.90396 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84386 +trainer/policy/normal/std Std 0.627749 +trainer/policy/normal/std Max 5.13673 +trainer/policy/normal/std Min 0.322406 +trainer/policy/normal/log_std Mean 1.01108 +trainer/policy/normal/log_std Std 0.292451 +trainer/policy/normal/log_std Max 1.63642 +trainer/policy/normal/log_std Min -1.13195 +eval/num steps total 367293 +eval/num paths total 369 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0261795 +eval/Actions Std 0.899159 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75385 +time/logging (s) 0.00407062 +time/sampling batch (s) 0.526839 +time/saving (s) 0.00396095 +time/training (s) 7.43037 +time/epoch (s) 10.7191 +time/total (s) 3845.28 +Epoch -632 +---------------------------------- --------------- +2022-05-10 14:14:56.031901 PDT | [1] Epoch -631 finished +---------------------------------- --------------- +epoch -631 +replay_buffer/size 999033 +trainer/num train calls 370000 +trainer/Policy Loss -19.4179 +trainer/Log Pis Mean 23.9402 +trainer/Log Pis Std 13.4299 +trainer/Log Pis Max 75.8459 +trainer/Log Pis Min -9.8505 +trainer/policy/mean Mean -0.0382781 +trainer/policy/mean Std 0.905217 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.87465 +trainer/policy/normal/std Std 0.619987 +trainer/policy/normal/std Max 5.42894 +trainer/policy/normal/std Min 0.307046 +trainer/policy/normal/log_std Mean 1.02348 +trainer/policy/normal/log_std Std 0.284849 +trainer/policy/normal/log_std Max 1.69174 +trainer/policy/normal/log_std Min -1.18076 +eval/num steps total 368293 +eval/num paths total 370 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.273591 +eval/Actions Std 0.906751 +eval/Actions Max 0.999985 +eval/Actions Min -0.999957 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46948 +time/logging (s) 0.00395805 +time/sampling batch (s) 0.274297 +time/saving (s) 0.00365615 +time/training (s) 8.0415 +time/epoch (s) 10.7929 +time/total (s) 3856.08 +Epoch -631 +---------------------------------- --------------- +2022-05-10 14:15:05.348301 PDT | [1] Epoch -630 finished +---------------------------------- --------------- +epoch -630 +replay_buffer/size 999033 +trainer/num train calls 371000 +trainer/Policy Loss -19.7147 +trainer/Log Pis Mean 23.3959 +trainer/Log Pis Std 12.5698 +trainer/Log Pis Max 59.9312 +trainer/Log Pis Min -10.3033 +trainer/policy/mean Mean -0.0239508 +trainer/policy/mean Std 0.903858 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80623 +trainer/policy/normal/std Std 0.619762 +trainer/policy/normal/std Max 4.91896 +trainer/policy/normal/std Min 0.38814 +trainer/policy/normal/log_std Mean 0.996899 +trainer/policy/normal/log_std Std 0.296672 +trainer/policy/normal/log_std Max 1.5931 +trainer/policy/normal/log_std Min -0.946389 +eval/num steps total 369293 +eval/num paths total 371 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.176432 +eval/Actions Std 0.915236 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40828 +time/logging (s) 0.00364197 +time/sampling batch (s) 0.277097 +time/saving (s) 0.00341826 +time/training (s) 6.60223 +time/epoch (s) 9.29467 +time/total (s) 3865.37 +Epoch -630 +---------------------------------- --------------- +2022-05-10 14:15:15.507610 PDT | [1] Epoch -629 finished +---------------------------------- --------------- +epoch -629 +replay_buffer/size 999033 +trainer/num train calls 372000 +trainer/Policy Loss -19.481 +trainer/Log Pis Mean 23.1103 +trainer/Log Pis Std 12.8147 +trainer/Log Pis Max 67.6006 +trainer/Log Pis Min -6.24575 +trainer/policy/mean Mean -0.0390989 +trainer/policy/mean Std 0.90328 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79214 +trainer/policy/normal/std Std 0.656697 +trainer/policy/normal/std Max 5.64687 +trainer/policy/normal/std Min 0.300687 +trainer/policy/normal/log_std Mean 0.986844 +trainer/policy/normal/log_std Std 0.319246 +trainer/policy/normal/log_std Max 1.7311 +trainer/policy/normal/log_std Min -1.20169 +eval/num steps total 370293 +eval/num paths total 372 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.394469 +eval/Actions Std 0.892343 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28875 +time/logging (s) 0.00409878 +time/sampling batch (s) 0.530241 +time/saving (s) 0.0039475 +time/training (s) 7.31042 +time/epoch (s) 10.1375 +time/total (s) 3875.51 +Epoch -629 +---------------------------------- --------------- +2022-05-10 14:15:26.938949 PDT | [1] Epoch -628 finished +---------------------------------- --------------- +epoch -628 +replay_buffer/size 999033 +trainer/num train calls 373000 +trainer/Policy Loss -19.1404 +trainer/Log Pis Mean 24.5689 +trainer/Log Pis Std 13.5479 +trainer/Log Pis Max 72.3039 +trainer/Log Pis Min -7.10616 +trainer/policy/mean Mean -0.0329881 +trainer/policy/mean Std 0.902742 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83844 +trainer/policy/normal/std Std 0.637628 +trainer/policy/normal/std Max 5.02572 +trainer/policy/normal/std Min 0.315363 +trainer/policy/normal/log_std Mean 1.00651 +trainer/policy/normal/log_std Std 0.306448 +trainer/policy/normal/log_std Max 1.61457 +trainer/policy/normal/log_std Min -1.15403 +eval/num steps total 371293 +eval/num paths total 373 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126248 +eval/Actions Std 0.903066 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66035 +time/logging (s) 0.00372129 +time/sampling batch (s) 0.527385 +time/saving (s) 0.00356769 +time/training (s) 8.21397 +time/epoch (s) 11.409 +time/total (s) 3886.93 +Epoch -628 +---------------------------------- --------------- +2022-05-10 14:15:36.094348 PDT | [1] Epoch -627 finished +---------------------------------- --------------- +epoch -627 +replay_buffer/size 999033 +trainer/num train calls 374000 +trainer/Policy Loss -20.1699 +trainer/Log Pis Mean 24.317 +trainer/Log Pis Std 14.0311 +trainer/Log Pis Max 66.3456 +trainer/Log Pis Min -13.6286 +trainer/policy/mean Mean -0.0359578 +trainer/policy/mean Std 0.9101 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81696 +trainer/policy/normal/std Std 0.628531 +trainer/policy/normal/std Max 5.42669 +trainer/policy/normal/std Min 0.309792 +trainer/policy/normal/log_std Mean 1.00029 +trainer/policy/normal/log_std Std 0.298657 +trainer/policy/normal/log_std Max 1.69133 +trainer/policy/normal/log_std Min -1.17186 +eval/num steps total 372293 +eval/num paths total 374 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196326 +eval/Actions Std 0.945191 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.20408 +time/logging (s) 0.00368135 +time/sampling batch (s) 0.277522 +time/saving (s) 0.00343852 +time/training (s) 6.64498 +time/epoch (s) 9.1337 +time/total (s) 3896.06 +Epoch -627 +---------------------------------- --------------- +2022-05-10 14:15:45.599421 PDT | [1] Epoch -626 finished +---------------------------------- --------------- +epoch -626 +replay_buffer/size 999033 +trainer/num train calls 375000 +trainer/Policy Loss -20.0506 +trainer/Log Pis Mean 24.3211 +trainer/Log Pis Std 13.5405 +trainer/Log Pis Max 72.5195 +trainer/Log Pis Min -11.2408 +trainer/policy/mean Mean -0.0350264 +trainer/policy/mean Std 0.91042 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79913 +trainer/policy/normal/std Std 0.620514 +trainer/policy/normal/std Max 5.66622 +trainer/policy/normal/std Min 0.369049 +trainer/policy/normal/log_std Mean 0.995241 +trainer/policy/normal/log_std Std 0.290522 +trainer/policy/normal/log_std Max 1.73452 +trainer/policy/normal/log_std Min -0.996826 +eval/num steps total 373293 +eval/num paths total 375 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.324599 +eval/Actions Std 0.878977 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50366 +time/logging (s) 0.00374515 +time/sampling batch (s) 0.275606 +time/saving (s) 0.00359117 +time/training (s) 6.69718 +time/epoch (s) 9.48379 +time/total (s) 3905.55 +Epoch -626 +---------------------------------- --------------- +2022-05-10 14:15:56.468028 PDT | [1] Epoch -625 finished +---------------------------------- --------------- +epoch -625 +replay_buffer/size 999033 +trainer/num train calls 376000 +trainer/Policy Loss -19.3684 +trainer/Log Pis Mean 23.6853 +trainer/Log Pis Std 13.0513 +trainer/Log Pis Max 62.7754 +trainer/Log Pis Min -5.86653 +trainer/policy/mean Mean -0.0335532 +trainer/policy/mean Std 0.908668 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.74568 +trainer/policy/normal/std Std 0.616788 +trainer/policy/normal/std Max 5.31736 +trainer/policy/normal/std Min 0.296435 +trainer/policy/normal/log_std Mean 0.973663 +trainer/policy/normal/log_std Std 0.303437 +trainer/policy/normal/log_std Max 1.67098 +trainer/policy/normal/log_std Min -1.21593 +eval/num steps total 374293 +eval/num paths total 376 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0777229 +eval/Actions Std 0.915007 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45162 +time/logging (s) 0.00369031 +time/sampling batch (s) 0.777391 +time/saving (s) 0.00342311 +time/training (s) 7.6105 +time/epoch (s) 10.8466 +time/total (s) 3916.4 +Epoch -625 +---------------------------------- --------------- +2022-05-10 14:16:06.159226 PDT | [1] Epoch -624 finished +---------------------------------- --------------- +epoch -624 +replay_buffer/size 999033 +trainer/num train calls 377000 +trainer/Policy Loss -19.0396 +trainer/Log Pis Mean 23.673 +trainer/Log Pis Std 12.4327 +trainer/Log Pis Max 66.1148 +trainer/Log Pis Min -6.68055 +trainer/policy/mean Mean -0.0245896 +trainer/policy/mean Std 0.903377 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76573 +trainer/policy/normal/std Std 0.648173 +trainer/policy/normal/std Max 7.14563 +trainer/policy/normal/std Min 0.286918 +trainer/policy/normal/log_std Mean 0.977903 +trainer/policy/normal/log_std Std 0.317612 +trainer/policy/normal/log_std Max 1.9665 +trainer/policy/normal/log_std Min -1.24856 +eval/num steps total 375293 +eval/num paths total 377 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0100883 +eval/Actions Std 0.967798 +eval/Actions Max 0.999997 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60267 +time/logging (s) 0.00414897 +time/sampling batch (s) 0.533632 +time/saving (s) 0.00382695 +time/training (s) 6.52527 +time/epoch (s) 9.66955 +time/total (s) 3926.07 +Epoch -624 +---------------------------------- --------------- +2022-05-10 14:16:16.547511 PDT | [1] Epoch -623 finished +---------------------------------- --------------- +epoch -623 +replay_buffer/size 999033 +trainer/num train calls 378000 +trainer/Policy Loss -19.1152 +trainer/Log Pis Mean 23.7804 +trainer/Log Pis Std 13.2434 +trainer/Log Pis Max 72.9116 +trainer/Log Pis Min -5.9446 +trainer/policy/mean Mean -0.0391623 +trainer/policy/mean Std 0.910096 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81495 +trainer/policy/normal/std Std 0.605932 +trainer/policy/normal/std Max 5.69735 +trainer/policy/normal/std Min 0.322653 +trainer/policy/normal/log_std Mean 1.00228 +trainer/policy/normal/log_std Std 0.286633 +trainer/policy/normal/log_std Max 1.74 +trainer/policy/normal/log_std Min -1.13118 +eval/num steps total 376293 +eval/num paths total 378 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.140271 +eval/Actions Std 0.871915 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.378 +time/logging (s) 0.00391038 +time/sampling batch (s) 0.284779 +time/saving (s) 0.00389693 +time/training (s) 7.69517 +time/epoch (s) 10.3658 +time/total (s) 3936.44 +Epoch -623 +---------------------------------- --------------- +2022-05-10 14:16:26.912716 PDT | [1] Epoch -622 finished +---------------------------------- --------------- +epoch -622 +replay_buffer/size 999033 +trainer/num train calls 379000 +trainer/Policy Loss -19.9624 +trainer/Log Pis Mean 24.7588 +trainer/Log Pis Std 13.7378 +trainer/Log Pis Max 70.971 +trainer/Log Pis Min -4.98017 +trainer/policy/mean Mean -0.0290547 +trainer/policy/mean Std 0.905862 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.84298 +trainer/policy/normal/std Std 0.605503 +trainer/policy/normal/std Max 5.73926 +trainer/policy/normal/std Min 0.376165 +trainer/policy/normal/log_std Mean 1.01309 +trainer/policy/normal/log_std Std 0.282201 +trainer/policy/normal/log_std Max 1.74733 +trainer/policy/normal/log_std Min -0.977726 +eval/num steps total 377293 +eval/num paths total 379 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0159524 +eval/Actions Std 0.907634 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5069 +time/logging (s) 0.00400127 +time/sampling batch (s) 0.53234 +time/saving (s) 0.00400253 +time/training (s) 7.29575 +time/epoch (s) 10.343 +time/total (s) 3946.79 +Epoch -622 +---------------------------------- --------------- +2022-05-10 14:16:38.028443 PDT | [1] Epoch -621 finished +---------------------------------- --------------- +epoch -621 +replay_buffer/size 999033 +trainer/num train calls 380000 +trainer/Policy Loss -20.0868 +trainer/Log Pis Mean 26.0783 +trainer/Log Pis Std 13.4967 +trainer/Log Pis Max 74.3096 +trainer/Log Pis Min -6.33342 +trainer/policy/mean Mean -0.0359517 +trainer/policy/mean Std 0.907431 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.84527 +trainer/policy/normal/std Std 0.641173 +trainer/policy/normal/std Max 5.61261 +trainer/policy/normal/std Min 0.382693 +trainer/policy/normal/log_std Mean 1.00956 +trainer/policy/normal/log_std Std 0.30082 +trainer/policy/normal/log_std Max 1.72502 +trainer/policy/normal/log_std Min -0.960521 +eval/num steps total 378293 +eval/num paths total 380 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104888 +eval/Actions Std 0.925484 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70763 +time/logging (s) 0.00404316 +time/sampling batch (s) 0.530151 +time/saving (s) 0.00386911 +time/training (s) 7.84785 +time/epoch (s) 11.0935 +time/total (s) 3957.88 +Epoch -621 +---------------------------------- --------------- +2022-05-10 14:16:47.611346 PDT | [1] Epoch -620 finished +---------------------------------- --------------- +epoch -620 +replay_buffer/size 999033 +trainer/num train calls 381000 +trainer/Policy Loss -19.4022 +trainer/Log Pis Mean 24.4521 +trainer/Log Pis Std 13.4749 +trainer/Log Pis Max 67.7874 +trainer/Log Pis Min -8.86863 +trainer/policy/mean Mean -0.0396724 +trainer/policy/mean Std 0.908462 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86918 +trainer/policy/normal/std Std 0.606075 +trainer/policy/normal/std Max 5.76769 +trainer/policy/normal/std Min 0.300611 +trainer/policy/normal/log_std Mean 1.02373 +trainer/policy/normal/log_std Std 0.273448 +trainer/policy/normal/log_std Max 1.75227 +trainer/policy/normal/log_std Min -1.20194 +eval/num steps total 379293 +eval/num paths total 381 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.120996 +eval/Actions Std 0.914094 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75723 +time/logging (s) 0.00380104 +time/sampling batch (s) 0.529802 +time/saving (s) 0.00377642 +time/training (s) 6.26567 +time/epoch (s) 9.56028 +time/total (s) 3967.45 +Epoch -620 +---------------------------------- --------------- +2022-05-10 14:16:57.150368 PDT | [1] Epoch -619 finished +---------------------------------- --------------- +epoch -619 +replay_buffer/size 999033 +trainer/num train calls 382000 +trainer/Policy Loss -20.4978 +trainer/Log Pis Mean 24.2785 +trainer/Log Pis Std 12.9298 +trainer/Log Pis Max 80.5326 +trainer/Log Pis Min -6.46339 +trainer/policy/mean Mean -0.0375482 +trainer/policy/mean Std 0.90677 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82193 +trainer/policy/normal/std Std 0.617008 +trainer/policy/normal/std Max 5.39022 +trainer/policy/normal/std Min 0.351542 +trainer/policy/normal/log_std Mean 1.00337 +trainer/policy/normal/log_std Std 0.293163 +trainer/policy/normal/log_std Max 1.68459 +trainer/policy/normal/log_std Min -1.04543 +eval/num steps total 380293 +eval/num paths total 382 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.149668 +eval/Actions Std 0.883617 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49817 +time/logging (s) 0.00439133 +time/sampling batch (s) 0.278546 +time/saving (s) 0.00409113 +time/training (s) 6.73251 +time/epoch (s) 9.51771 +time/total (s) 3976.97 +Epoch -619 +---------------------------------- --------------- +2022-05-10 14:17:07.534635 PDT | [1] Epoch -618 finished +---------------------------------- --------------- +epoch -618 +replay_buffer/size 999033 +trainer/num train calls 383000 +trainer/Policy Loss -19.8404 +trainer/Log Pis Mean 23.7843 +trainer/Log Pis Std 14.1518 +trainer/Log Pis Max 75.7222 +trainer/Log Pis Min -10.2118 +trainer/policy/mean Mean -0.0174407 +trainer/policy/mean Std 0.905068 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82191 +trainer/policy/normal/std Std 0.655272 +trainer/policy/normal/std Max 6.19666 +trainer/policy/normal/std Min 0.317644 +trainer/policy/normal/log_std Mean 0.999489 +trainer/policy/normal/log_std Std 0.308753 +trainer/policy/normal/log_std Max 1.82401 +trainer/policy/normal/log_std Min -1.14682 +eval/num steps total 381293 +eval/num paths total 383 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110229 +eval/Actions Std 0.90424 +eval/Actions Max 0.99998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49375 +time/logging (s) 0.0036819 +time/sampling batch (s) 0.525096 +time/saving (s) 0.00355157 +time/training (s) 7.33548 +time/epoch (s) 10.3616 +time/total (s) 3987.33 +Epoch -618 +---------------------------------- --------------- +2022-05-10 14:17:18.053192 PDT | [1] Epoch -617 finished +---------------------------------- --------------- +epoch -617 +replay_buffer/size 999033 +trainer/num train calls 384000 +trainer/Policy Loss -20.0778 +trainer/Log Pis Mean 24.021 +trainer/Log Pis Std 13.2343 +trainer/Log Pis Max 77.1566 +trainer/Log Pis Min -5.01963 +trainer/policy/mean Mean -0.0398332 +trainer/policy/mean Std 0.907077 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75115 +trainer/policy/normal/std Std 0.606574 +trainer/policy/normal/std Max 5.52393 +trainer/policy/normal/std Min 0.327363 +trainer/policy/normal/log_std Mean 0.977465 +trainer/policy/normal/log_std Std 0.294989 +trainer/policy/normal/log_std Max 1.70909 +trainer/policy/normal/log_std Min -1.11669 +eval/num steps total 382293 +eval/num paths total 384 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0011644 +eval/Actions Std 0.8877 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58623 +time/logging (s) 0.00405552 +time/sampling batch (s) 1.02572 +time/saving (s) 0.00381015 +time/training (s) 6.87751 +time/epoch (s) 10.4973 +time/total (s) 3997.83 +Epoch -617 +---------------------------------- --------------- +2022-05-10 14:17:27.977787 PDT | [1] Epoch -616 finished +---------------------------------- --------------- +epoch -616 +replay_buffer/size 999033 +trainer/num train calls 385000 +trainer/Policy Loss -19.6143 +trainer/Log Pis Mean 24.4511 +trainer/Log Pis Std 13.3717 +trainer/Log Pis Max 75.0013 +trainer/Log Pis Min -5.25193 +trainer/policy/mean Mean -0.0347522 +trainer/policy/mean Std 0.907191 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84821 +trainer/policy/normal/std Std 0.606701 +trainer/policy/normal/std Max 4.96373 +trainer/policy/normal/std Min 0.31421 +trainer/policy/normal/log_std Mean 1.0154 +trainer/policy/normal/log_std Std 0.278406 +trainer/policy/normal/log_std Max 1.60216 +trainer/policy/normal/log_std Min -1.15769 +eval/num steps total 383293 +eval/num paths total 385 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0255789 +eval/Actions Std 0.905084 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66049 +time/logging (s) 0.00385691 +time/sampling batch (s) 0.278756 +time/saving (s) 0.00378505 +time/training (s) 6.95539 +time/epoch (s) 9.90227 +time/total (s) 4007.74 +Epoch -616 +---------------------------------- --------------- +2022-05-10 14:17:37.593766 PDT | [1] Epoch -615 finished +---------------------------------- --------------- +epoch -615 +replay_buffer/size 999033 +trainer/num train calls 386000 +trainer/Policy Loss -20.2381 +trainer/Log Pis Mean 25.522 +trainer/Log Pis Std 13.7569 +trainer/Log Pis Max 72.8163 +trainer/Log Pis Min -12.7702 +trainer/policy/mean Mean -0.0512552 +trainer/policy/mean Std 0.908932 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8206 +trainer/policy/normal/std Std 0.642101 +trainer/policy/normal/std Max 6.26158 +trainer/policy/normal/std Min 0.311347 +trainer/policy/normal/log_std Mean 0.999401 +trainer/policy/normal/log_std Std 0.309921 +trainer/policy/normal/log_std Max 1.83443 +trainer/policy/normal/log_std Min -1.16685 +eval/num steps total 384293 +eval/num paths total 386 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0791253 +eval/Actions Std 0.911345 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64118 +time/logging (s) 0.00367357 +time/sampling batch (s) 0.277163 +time/saving (s) 0.00339521 +time/training (s) 6.66849 +time/epoch (s) 9.5939 +time/total (s) 4017.33 +Epoch -615 +---------------------------------- --------------- +2022-05-10 14:17:47.292877 PDT | [1] Epoch -614 finished +---------------------------------- --------------- +epoch -614 +replay_buffer/size 999033 +trainer/num train calls 387000 +trainer/Policy Loss -18.7003 +trainer/Log Pis Mean 25.2029 +trainer/Log Pis Std 13.8124 +trainer/Log Pis Max 76.9154 +trainer/Log Pis Min -5.16281 +trainer/policy/mean Mean -0.020494 +trainer/policy/mean Std 0.903902 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85847 +trainer/policy/normal/std Std 0.642743 +trainer/policy/normal/std Max 5.45358 +trainer/policy/normal/std Min 0.323492 +trainer/policy/normal/log_std Mean 1.0129 +trainer/policy/normal/log_std Std 0.310762 +trainer/policy/normal/log_std Max 1.69627 +trainer/policy/normal/log_std Min -1.12858 +eval/num steps total 385293 +eval/num paths total 387 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12494 +eval/Actions Std 0.901803 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59363 +time/logging (s) 0.00406488 +time/sampling batch (s) 0.524781 +time/saving (s) 0.00374344 +time/training (s) 6.5518 +time/epoch (s) 9.67803 +time/total (s) 4027.02 +Epoch -614 +---------------------------------- --------------- +2022-05-10 14:17:56.392517 PDT | [1] Epoch -613 finished +---------------------------------- --------------- +epoch -613 +replay_buffer/size 999033 +trainer/num train calls 388000 +trainer/Policy Loss -20.4837 +trainer/Log Pis Mean 24.3312 +trainer/Log Pis Std 13.4898 +trainer/Log Pis Max 71.3711 +trainer/Log Pis Min -8.8836 +trainer/policy/mean Mean -0.0548086 +trainer/policy/mean Std 0.909591 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85812 +trainer/policy/normal/std Std 0.62663 +trainer/policy/normal/std Max 5.67122 +trainer/policy/normal/std Min 0.339233 +trainer/policy/normal/log_std Mean 1.01636 +trainer/policy/normal/log_std Std 0.291642 +trainer/policy/normal/log_std Max 1.7354 +trainer/policy/normal/log_std Min -1.08107 +eval/num steps total 386293 +eval/num paths total 388 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.305987 +eval/Actions Std 0.907232 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68224 +time/logging (s) 0.00391372 +time/sampling batch (s) 0.277976 +time/saving (s) 0.00371371 +time/training (s) 6.10925 +time/epoch (s) 9.07709 +time/total (s) 4036.1 +Epoch -613 +---------------------------------- --------------- +2022-05-10 14:18:07.739826 PDT | [1] Epoch -612 finished +---------------------------------- --------------- +epoch -612 +replay_buffer/size 999033 +trainer/num train calls 389000 +trainer/Policy Loss -19.5854 +trainer/Log Pis Mean 24.1079 +trainer/Log Pis Std 13.0372 +trainer/Log Pis Max 62.468 +trainer/Log Pis Min -8.29886 +trainer/policy/mean Mean -0.0412164 +trainer/policy/mean Std 0.908212 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79656 +trainer/policy/normal/std Std 0.639366 +trainer/policy/normal/std Max 6.37507 +trainer/policy/normal/std Min 0.340353 +trainer/policy/normal/log_std Mean 0.990766 +trainer/policy/normal/log_std Std 0.309471 +trainer/policy/normal/log_std Max 1.8524 +trainer/policy/normal/log_std Min -1.07777 +eval/num steps total 387293 +eval/num paths total 389 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.1498 +eval/Actions Std 0.864165 +eval/Actions Max 0.999993 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74757 +time/logging (s) 0.00376896 +time/sampling batch (s) 0.286549 +time/saving (s) 0.00344052 +time/training (s) 8.28356 +time/epoch (s) 11.3249 +time/total (s) 4047.42 +Epoch -612 +---------------------------------- --------------- +2022-05-10 14:18:19.420805 PDT | [1] Epoch -611 finished +---------------------------------- --------------- +epoch -611 +replay_buffer/size 999033 +trainer/num train calls 390000 +trainer/Policy Loss -20.4687 +trainer/Log Pis Mean 25.3989 +trainer/Log Pis Std 13.8992 +trainer/Log Pis Max 70.4326 +trainer/Log Pis Min -6.27202 +trainer/policy/mean Mean -0.0315951 +trainer/policy/mean Std 0.911621 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81942 +trainer/policy/normal/std Std 0.642617 +trainer/policy/normal/std Max 6.01022 +trainer/policy/normal/std Min 0.33466 +trainer/policy/normal/log_std Mean 0.998318 +trainer/policy/normal/log_std Std 0.313672 +trainer/policy/normal/log_std Max 1.79346 +trainer/policy/normal/log_std Min -1.09464 +eval/num steps total 388293 +eval/num paths total 390 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.020769 +eval/Actions Std 0.872183 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53174 +time/logging (s) 0.00379153 +time/sampling batch (s) 1.28388 +time/saving (s) 0.00343517 +time/training (s) 7.83609 +time/epoch (s) 11.6589 +time/total (s) 4059.09 +Epoch -611 +---------------------------------- --------------- +2022-05-10 14:18:29.518668 PDT | [1] Epoch -610 finished +---------------------------------- --------------- +epoch -610 +replay_buffer/size 999033 +trainer/num train calls 391000 +trainer/Policy Loss -19.8136 +trainer/Log Pis Mean 24.9713 +trainer/Log Pis Std 12.8649 +trainer/Log Pis Max 67.3193 +trainer/Log Pis Min -8.32847 +trainer/policy/mean Mean -0.00908223 +trainer/policy/mean Std 0.91066 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80007 +trainer/policy/normal/std Std 0.605088 +trainer/policy/normal/std Max 5.31201 +trainer/policy/normal/std Min 0.320799 +trainer/policy/normal/log_std Mean 0.9961 +trainer/policy/normal/log_std Std 0.291535 +trainer/policy/normal/log_std Max 1.66997 +trainer/policy/normal/log_std Min -1.13694 +eval/num steps total 389293 +eval/num paths total 391 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.295036 +eval/Actions Std 0.88186 +eval/Actions Max 0.999983 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60747 +time/logging (s) 0.0039527 +time/sampling batch (s) 0.53101 +time/saving (s) 0.00401283 +time/training (s) 6.92953 +time/epoch (s) 10.076 +time/total (s) 4069.16 +Epoch -610 +---------------------------------- --------------- +2022-05-10 14:18:39.998264 PDT | [1] Epoch -609 finished +---------------------------------- --------------- +epoch -609 +replay_buffer/size 999033 +trainer/num train calls 392000 +trainer/Policy Loss -19.321 +trainer/Log Pis Mean 23.9797 +trainer/Log Pis Std 12.8527 +trainer/Log Pis Max 76.5851 +trainer/Log Pis Min -3.36679 +trainer/policy/mean Mean -0.0111469 +trainer/policy/mean Std 0.905827 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77074 +trainer/policy/normal/std Std 0.621675 +trainer/policy/normal/std Max 5.0539 +trainer/policy/normal/std Min 0.348294 +trainer/policy/normal/log_std Mean 0.983204 +trainer/policy/normal/log_std Std 0.300938 +trainer/policy/normal/log_std Max 1.62016 +trainer/policy/normal/log_std Min -1.05471 +eval/num steps total 390293 +eval/num paths total 392 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0700601 +eval/Actions Std 0.894069 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71174 +time/logging (s) 0.00368456 +time/sampling batch (s) 0.280034 +time/saving (s) 0.00345395 +time/training (s) 7.45804 +time/epoch (s) 10.457 +time/total (s) 4079.62 +Epoch -609 +---------------------------------- --------------- +2022-05-10 14:18:51.665175 PDT | [1] Epoch -608 finished +---------------------------------- --------------- +epoch -608 +replay_buffer/size 999033 +trainer/num train calls 393000 +trainer/Policy Loss -19.3223 +trainer/Log Pis Mean 24.2881 +trainer/Log Pis Std 12.8051 +trainer/Log Pis Max 60.2804 +trainer/Log Pis Min -4.28078 +trainer/policy/mean Mean -0.0356342 +trainer/policy/mean Std 0.901822 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.79875 +trainer/policy/normal/std Std 0.634274 +trainer/policy/normal/std Max 5.37157 +trainer/policy/normal/std Min 0.3021 +trainer/policy/normal/log_std Mean 0.990581 +trainer/policy/normal/log_std Std 0.31657 +trainer/policy/normal/log_std Max 1.68112 +trainer/policy/normal/log_std Min -1.197 +eval/num steps total 391293 +eval/num paths total 393 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0770626 +eval/Actions Std 0.912385 +eval/Actions Max 0.999991 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58124 +time/logging (s) 0.00428752 +time/sampling batch (s) 0.781134 +time/saving (s) 0.00416181 +time/training (s) 8.27488 +time/epoch (s) 11.6457 +time/total (s) 4091.27 +Epoch -608 +---------------------------------- --------------- +2022-05-10 14:19:01.516768 PDT | [1] Epoch -607 finished +---------------------------------- --------------- +epoch -607 +replay_buffer/size 999033 +trainer/num train calls 394000 +trainer/Policy Loss -19.689 +trainer/Log Pis Mean 24.7826 +trainer/Log Pis Std 13.0563 +trainer/Log Pis Max 66.0851 +trainer/Log Pis Min -5.45799 +trainer/policy/mean Mean -0.0245531 +trainer/policy/mean Std 0.902316 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83271 +trainer/policy/normal/std Std 0.62766 +trainer/policy/normal/std Max 5.34881 +trainer/policy/normal/std Min 0.343484 +trainer/policy/normal/log_std Mean 1.00556 +trainer/policy/normal/log_std Std 0.302479 +trainer/policy/normal/log_std Max 1.67687 +trainer/policy/normal/log_std Min -1.06861 +eval/num steps total 392293 +eval/num paths total 394 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110635 +eval/Actions Std 0.913269 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64588 +time/logging (s) 0.00402899 +time/sampling batch (s) 0.28129 +time/saving (s) 0.00397234 +time/training (s) 6.8935 +time/epoch (s) 9.82867 +time/total (s) 4101.11 +Epoch -607 +---------------------------------- --------------- +2022-05-10 14:19:12.181192 PDT | [1] Epoch -606 finished +---------------------------------- --------------- +epoch -606 +replay_buffer/size 999033 +trainer/num train calls 395000 +trainer/Policy Loss -20.3349 +trainer/Log Pis Mean 23.4453 +trainer/Log Pis Std 12.4601 +trainer/Log Pis Max 62.2039 +trainer/Log Pis Min -4.45095 +trainer/policy/mean Mean -0.0448803 +trainer/policy/mean Std 0.902553 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78427 +trainer/policy/normal/std Std 0.606756 +trainer/policy/normal/std Max 5.33476 +trainer/policy/normal/std Min 0.365385 +trainer/policy/normal/log_std Mean 0.990295 +trainer/policy/normal/log_std Std 0.290799 +trainer/policy/normal/log_std Max 1.67424 +trainer/policy/normal/log_std Min -1.0068 +eval/num steps total 393293 +eval/num paths total 395 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249264 +eval/Actions Std 0.873657 +eval/Actions Max 0.999988 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61518 +time/logging (s) 0.00411417 +time/sampling batch (s) 0.782519 +time/saving (s) 0.00402501 +time/training (s) 7.23597 +time/epoch (s) 10.6418 +time/total (s) 4111.75 +Epoch -606 +---------------------------------- --------------- +2022-05-10 14:19:22.780713 PDT | [1] Epoch -605 finished +---------------------------------- --------------- +epoch -605 +replay_buffer/size 999033 +trainer/num train calls 396000 +trainer/Policy Loss -18.8621 +trainer/Log Pis Mean 24.3868 +trainer/Log Pis Std 12.8549 +trainer/Log Pis Max 79.4496 +trainer/Log Pis Min -5.20687 +trainer/policy/mean Mean -0.0370271 +trainer/policy/mean Std 0.904384 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83617 +trainer/policy/normal/std Std 0.625504 +trainer/policy/normal/std Max 5.05337 +trainer/policy/normal/std Min 0.274141 +trainer/policy/normal/log_std Mean 1.00713 +trainer/policy/normal/log_std Std 0.301063 +trainer/policy/normal/log_std Max 1.62006 +trainer/policy/normal/log_std Min -1.29411 +eval/num steps total 394293 +eval/num paths total 396 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.149654 +eval/Actions Std 0.91238 +eval/Actions Max 0.999981 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55055 +time/logging (s) 0.00427898 +time/sampling batch (s) 0.282358 +time/saving (s) 0.00413077 +time/training (s) 7.73569 +time/epoch (s) 10.577 +time/total (s) 4122.33 +Epoch -605 +---------------------------------- --------------- +2022-05-10 14:19:33.609806 PDT | [1] Epoch -604 finished +---------------------------------- --------------- +epoch -604 +replay_buffer/size 999033 +trainer/num train calls 397000 +trainer/Policy Loss -19.4652 +trainer/Log Pis Mean 24.0869 +trainer/Log Pis Std 13.3769 +trainer/Log Pis Max 62.539 +trainer/Log Pis Min -9.24389 +trainer/policy/mean Mean -0.0190209 +trainer/policy/mean Std 0.908 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81515 +trainer/policy/normal/std Std 0.609396 +trainer/policy/normal/std Max 5.68157 +trainer/policy/normal/std Min 0.375128 +trainer/policy/normal/log_std Mean 1.00302 +trainer/policy/normal/log_std Std 0.280267 +trainer/policy/normal/log_std Max 1.73723 +trainer/policy/normal/log_std Min -0.980488 +eval/num steps total 395293 +eval/num paths total 397 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0220524 +eval/Actions Std 0.910942 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48709 +time/logging (s) 0.00372946 +time/sampling batch (s) 0.286087 +time/saving (s) 0.00348446 +time/training (s) 8.02518 +time/epoch (s) 10.8056 +time/total (s) 4133.14 +Epoch -604 +---------------------------------- --------------- +2022-05-10 14:19:43.702584 PDT | [1] Epoch -603 finished +---------------------------------- --------------- +epoch -603 +replay_buffer/size 999033 +trainer/num train calls 398000 +trainer/Policy Loss -19.1786 +trainer/Log Pis Mean 24.8251 +trainer/Log Pis Std 13.471 +trainer/Log Pis Max 73.5043 +trainer/Log Pis Min -9.19097 +trainer/policy/mean Mean -0.0718102 +trainer/policy/mean Std 0.91034 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.83742 +trainer/policy/normal/std Std 0.650177 +trainer/policy/normal/std Max 7.2361 +trainer/policy/normal/std Min 0.308235 +trainer/policy/normal/log_std Mean 1.00474 +trainer/policy/normal/log_std Std 0.313059 +trainer/policy/normal/log_std Max 1.97908 +trainer/policy/normal/log_std Min -1.17689 +eval/num steps total 396293 +eval/num paths total 398 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132371 +eval/Actions Std 0.884199 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66915 +time/logging (s) 0.00372157 +time/sampling batch (s) 0.281697 +time/saving (s) 0.00345977 +time/training (s) 7.11252 +time/epoch (s) 10.0705 +time/total (s) 4143.21 +Epoch -603 +---------------------------------- --------------- +2022-05-10 14:19:54.110808 PDT | [1] Epoch -602 finished +---------------------------------- --------------- +epoch -602 +replay_buffer/size 999033 +trainer/num train calls 399000 +trainer/Policy Loss -19.8971 +trainer/Log Pis Mean 24.0513 +trainer/Log Pis Std 13.2189 +trainer/Log Pis Max 68.7118 +trainer/Log Pis Min -10.7198 +trainer/policy/mean Mean -0.0287884 +trainer/policy/mean Std 0.907105 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.86444 +trainer/policy/normal/std Std 0.656185 +trainer/policy/normal/std Max 5.54015 +trainer/policy/normal/std Min 0.367454 +trainer/policy/normal/log_std Mean 1.01477 +trainer/policy/normal/log_std Std 0.308634 +trainer/policy/normal/log_std Max 1.71202 +trainer/policy/normal/log_std Min -1.00116 +eval/num steps total 397293 +eval/num paths total 399 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0755458 +eval/Actions Std 0.917996 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61832 +time/logging (s) 0.00399714 +time/sampling batch (s) 0.282429 +time/saving (s) 0.00402421 +time/training (s) 7.47741 +time/epoch (s) 10.3862 +time/total (s) 4153.6 +Epoch -602 +---------------------------------- --------------- +2022-05-10 14:20:04.500197 PDT | [1] Epoch -601 finished +---------------------------------- --------------- +epoch -601 +replay_buffer/size 999033 +trainer/num train calls 400000 +trainer/Policy Loss -19.0626 +trainer/Log Pis Mean 23.9901 +trainer/Log Pis Std 13.1383 +trainer/Log Pis Max 80.4404 +trainer/Log Pis Min -3.98915 +trainer/policy/mean Mean -0.0413911 +trainer/policy/mean Std 0.906551 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.78574 +trainer/policy/normal/std Std 0.622486 +trainer/policy/normal/std Max 5.73331 +trainer/policy/normal/std Min 0.320405 +trainer/policy/normal/log_std Mean 0.988897 +trainer/policy/normal/log_std Std 0.299924 +trainer/policy/normal/log_std Max 1.74629 +trainer/policy/normal/log_std Min -1.13817 +eval/num steps total 398293 +eval/num paths total 400 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0869222 +eval/Actions Std 0.887652 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71724 +time/logging (s) 0.00369606 +time/sampling batch (s) 0.288814 +time/saving (s) 0.00648616 +time/training (s) 7.34999 +time/epoch (s) 10.3662 +time/total (s) 4163.97 +Epoch -601 +---------------------------------- --------------- +2022-05-10 14:20:15.307467 PDT | [1] Epoch -600 finished +---------------------------------- --------------- +epoch -600 +replay_buffer/size 999033 +trainer/num train calls 401000 +trainer/Policy Loss -19.6958 +trainer/Log Pis Mean 23.9987 +trainer/Log Pis Std 13.2953 +trainer/Log Pis Max 76.7859 +trainer/Log Pis Min -11.0059 +trainer/policy/mean Mean -0.0337039 +trainer/policy/mean Std 0.909318 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84236 +trainer/policy/normal/std Std 0.63516 +trainer/policy/normal/std Max 5.53883 +trainer/policy/normal/std Min 0.269644 +trainer/policy/normal/log_std Mean 1.00878 +trainer/policy/normal/log_std Std 0.302035 +trainer/policy/normal/log_std Max 1.71178 +trainer/policy/normal/log_std Min -1.31065 +eval/num steps total 399293 +eval/num paths total 401 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.331079 +eval/Actions Std 0.863567 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58144 +time/logging (s) 0.00372593 +time/sampling batch (s) 0.283933 +time/saving (s) 0.00355481 +time/training (s) 7.91234 +time/epoch (s) 10.785 +time/total (s) 4174.76 +Epoch -600 +---------------------------------- --------------- +2022-05-10 14:20:25.385501 PDT | [1] Epoch -599 finished +---------------------------------- --------------- +epoch -599 +replay_buffer/size 999033 +trainer/num train calls 402000 +trainer/Policy Loss -20.1336 +trainer/Log Pis Mean 25.6857 +trainer/Log Pis Std 14.1157 +trainer/Log Pis Max 83.1089 +trainer/Log Pis Min -12.6969 +trainer/policy/mean Mean -0.0534 +trainer/policy/mean Std 0.9045 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81731 +trainer/policy/normal/std Std 0.638926 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.315709 +trainer/policy/normal/log_std Mean 0.999275 +trainer/policy/normal/log_std Std 0.304227 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.15293 +eval/num steps total 400293 +eval/num paths total 402 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0819297 +eval/Actions Std 0.915153 +eval/Actions Max 0.999991 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64792 +time/logging (s) 0.00377171 +time/sampling batch (s) 0.282844 +time/saving (s) 0.00346009 +time/training (s) 6.86716 +time/epoch (s) 9.80515 +time/total (s) 4184.82 +Epoch -599 +---------------------------------- --------------- +2022-05-10 14:20:35.323065 PDT | [1] Epoch -598 finished +---------------------------------- --------------- +epoch -598 +replay_buffer/size 999033 +trainer/num train calls 403000 +trainer/Policy Loss -19.3276 +trainer/Log Pis Mean 24.7298 +trainer/Log Pis Std 13.5444 +trainer/Log Pis Max 71.4641 +trainer/Log Pis Min -7.2066 +trainer/policy/mean Mean -0.0321378 +trainer/policy/mean Std 0.902697 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.88337 +trainer/policy/normal/std Std 0.628257 +trainer/policy/normal/std Max 6.09393 +trainer/policy/normal/std Min 0.305648 +trainer/policy/normal/log_std Mean 1.02544 +trainer/policy/normal/log_std Std 0.29167 +trainer/policy/normal/log_std Max 1.80729 +trainer/policy/normal/log_std Min -1.18532 +eval/num steps total 401293 +eval/num paths total 403 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.21598 +eval/Actions Std 0.926086 +eval/Actions Max 0.999984 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69159 +time/logging (s) 0.00403505 +time/sampling batch (s) 0.282645 +time/saving (s) 0.00407858 +time/training (s) 6.9332 +time/epoch (s) 9.91555 +time/total (s) 4194.74 +Epoch -598 +---------------------------------- --------------- +2022-05-10 14:20:46.300115 PDT | [1] Epoch -597 finished +---------------------------------- --------------- +epoch -597 +replay_buffer/size 999033 +trainer/num train calls 404000 +trainer/Policy Loss -17.9608 +trainer/Log Pis Mean 24.0594 +trainer/Log Pis Std 13.543 +trainer/Log Pis Max 62.4812 +trainer/Log Pis Min -9.38666 +trainer/policy/mean Mean -0.0505196 +trainer/policy/mean Std 0.905812 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83505 +trainer/policy/normal/std Std 0.621225 +trainer/policy/normal/std Max 5.40719 +trainer/policy/normal/std Min 0.326626 +trainer/policy/normal/log_std Mean 1.00876 +trainer/policy/normal/log_std Std 0.287924 +trainer/policy/normal/log_std Max 1.68773 +trainer/policy/normal/log_std Min -1.11894 +eval/num steps total 402293 +eval/num paths total 404 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222443 +eval/Actions Std 0.867644 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63103 +time/logging (s) 0.00377369 +time/sampling batch (s) 0.291334 +time/saving (s) 0.003579 +time/training (s) 8.02399 +time/epoch (s) 10.9537 +time/total (s) 4205.69 +Epoch -597 +---------------------------------- --------------- +2022-05-10 14:20:56.260898 PDT | [1] Epoch -596 finished +---------------------------------- --------------- +epoch -596 +replay_buffer/size 999033 +trainer/num train calls 405000 +trainer/Policy Loss -19.2326 +trainer/Log Pis Mean 24.6368 +trainer/Log Pis Std 13.5141 +trainer/Log Pis Max 68.892 +trainer/Log Pis Min -7.09101 +trainer/policy/mean Mean -0.0464843 +trainer/policy/mean Std 0.900455 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.83978 +trainer/policy/normal/std Std 0.62656 +trainer/policy/normal/std Max 6.55006 +trainer/policy/normal/std Min 0.288549 +trainer/policy/normal/log_std Mean 1.00848 +trainer/policy/normal/log_std Std 0.300002 +trainer/policy/normal/log_std Max 1.87947 +trainer/policy/normal/log_std Min -1.24289 +eval/num steps total 403293 +eval/num paths total 405 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.322202 +eval/Actions Std 0.794171 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6069 +time/logging (s) 0.00380261 +time/sampling batch (s) 0.28697 +time/saving (s) 0.00346346 +time/training (s) 7.0374 +time/epoch (s) 9.93854 +time/total (s) 4215.64 +Epoch -596 +---------------------------------- --------------- +2022-05-10 14:21:07.380254 PDT | [1] Epoch -595 finished +---------------------------------- --------------- +epoch -595 +replay_buffer/size 999033 +trainer/num train calls 406000 +trainer/Policy Loss -20.0562 +trainer/Log Pis Mean 24.3963 +trainer/Log Pis Std 13.216 +trainer/Log Pis Max 72.0568 +trainer/Log Pis Min -9.28732 +trainer/policy/mean Mean -0.0284443 +trainer/policy/mean Std 0.907357 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81757 +trainer/policy/normal/std Std 0.605231 +trainer/policy/normal/std Max 6.323 +trainer/policy/normal/std Min 0.27879 +trainer/policy/normal/log_std Mean 1.00318 +trainer/policy/normal/log_std Std 0.287929 +trainer/policy/normal/log_std Max 1.84419 +trainer/policy/normal/log_std Min -1.2773 +eval/num steps total 404293 +eval/num paths total 406 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.341236 +eval/Actions Std 0.835067 +eval/Actions Max 0.999995 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60987 +time/logging (s) 0.00444669 +time/sampling batch (s) 0.281393 +time/saving (s) 0.00418691 +time/training (s) 8.1977 +time/epoch (s) 11.0976 +time/total (s) 4226.74 +Epoch -595 +---------------------------------- --------------- +2022-05-10 14:21:17.872060 PDT | [1] Epoch -594 finished +---------------------------------- --------------- +epoch -594 +replay_buffer/size 999033 +trainer/num train calls 407000 +trainer/Policy Loss -19.1742 +trainer/Log Pis Mean 24.4451 +trainer/Log Pis Std 14.1076 +trainer/Log Pis Max 66.0994 +trainer/Log Pis Min -7.84809 +trainer/policy/mean Mean -0.0428462 +trainer/policy/mean Std 0.908595 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79617 +trainer/policy/normal/std Std 0.644329 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.248029 +trainer/policy/normal/log_std Mean 0.989979 +trainer/policy/normal/log_std Std 0.312138 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.39421 +eval/num steps total 405293 +eval/num paths total 407 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102302 +eval/Actions Std 0.932232 +eval/Actions Max 0.999985 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62558 +time/logging (s) 0.00375846 +time/sampling batch (s) 0.2791 +time/saving (s) 0.00342796 +time/training (s) 7.55653 +time/epoch (s) 10.4684 +time/total (s) 4237.21 +Epoch -594 +---------------------------------- --------------- +2022-05-10 14:21:27.673792 PDT | [1] Epoch -593 finished +---------------------------------- --------------- +epoch -593 +replay_buffer/size 999033 +trainer/num train calls 408000 +trainer/Policy Loss -18.9098 +trainer/Log Pis Mean 23.0289 +trainer/Log Pis Std 12.5495 +trainer/Log Pis Max 67.9208 +trainer/Log Pis Min -13.8609 +trainer/policy/mean Mean -0.0327685 +trainer/policy/mean Std 0.899203 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.74298 +trainer/policy/normal/std Std 0.645566 +trainer/policy/normal/std Max 5.83499 +trainer/policy/normal/std Min 0.336421 +trainer/policy/normal/log_std Mean 0.969924 +trainer/policy/normal/log_std Std 0.313996 +trainer/policy/normal/log_std Max 1.76387 +trainer/policy/normal/log_std Min -1.08939 +eval/num steps total 406293 +eval/num paths total 408 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.162766 +eval/Actions Std 0.87175 +eval/Actions Max 0.999985 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74691 +time/logging (s) 0.00380413 +time/sampling batch (s) 0.286117 +time/saving (s) 0.00407868 +time/training (s) 6.73835 +time/epoch (s) 9.77926 +time/total (s) 4246.99 +Epoch -593 +---------------------------------- --------------- +2022-05-10 14:21:38.006897 PDT | [1] Epoch -592 finished +---------------------------------- --------------- +epoch -592 +replay_buffer/size 999033 +trainer/num train calls 409000 +trainer/Policy Loss -19.8897 +trainer/Log Pis Mean 24.4759 +trainer/Log Pis Std 13.3305 +trainer/Log Pis Max 76.9067 +trainer/Log Pis Min -8.16954 +trainer/policy/mean Mean -0.0456141 +trainer/policy/mean Std 0.904405 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86928 +trainer/policy/normal/std Std 0.648111 +trainer/policy/normal/std Max 6.09633 +trainer/policy/normal/std Min 0.383489 +trainer/policy/normal/log_std Mean 1.01795 +trainer/policy/normal/log_std Std 0.302295 +trainer/policy/normal/log_std Max 1.80769 +trainer/policy/normal/log_std Min -0.958444 +eval/num steps total 407293 +eval/num paths total 409 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122758 +eval/Actions Std 0.906284 +eval/Actions Max 0.999987 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65777 +time/logging (s) 0.00417393 +time/sampling batch (s) 0.532965 +time/saving (s) 0.00392139 +time/training (s) 7.11153 +time/epoch (s) 10.3104 +time/total (s) 4257.3 +Epoch -592 +---------------------------------- --------------- +2022-05-10 14:21:48.085904 PDT | [1] Epoch -591 finished +---------------------------------- --------------- +epoch -591 +replay_buffer/size 999033 +trainer/num train calls 410000 +trainer/Policy Loss -19.8899 +trainer/Log Pis Mean 24.5554 +trainer/Log Pis Std 13.9957 +trainer/Log Pis Max 70.4387 +trainer/Log Pis Min -6.18041 +trainer/policy/mean Mean -0.0287288 +trainer/policy/mean Std 0.906742 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84878 +trainer/policy/normal/std Std 0.646861 +trainer/policy/normal/std Max 5.73569 +trainer/policy/normal/std Min 0.327176 +trainer/policy/normal/log_std Mean 1.01057 +trainer/policy/normal/log_std Std 0.30237 +trainer/policy/normal/log_std Max 1.74671 +trainer/policy/normal/log_std Min -1.11726 +eval/num steps total 408293 +eval/num paths total 410 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121216 +eval/Actions Std 0.90759 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46041 +time/logging (s) 0.00394856 +time/sampling batch (s) 0.779162 +time/saving (s) 0.0035248 +time/training (s) 6.80932 +time/epoch (s) 10.0564 +time/total (s) 4267.36 +Epoch -591 +---------------------------------- --------------- +2022-05-10 14:21:59.006734 PDT | [1] Epoch -590 finished +---------------------------------- --------------- +epoch -590 +replay_buffer/size 999033 +trainer/num train calls 411000 +trainer/Policy Loss -18.5962 +trainer/Log Pis Mean 23.5228 +trainer/Log Pis Std 13.3325 +trainer/Log Pis Max 61.6606 +trainer/Log Pis Min -6.84484 +trainer/policy/mean Mean -0.0339567 +trainer/policy/mean Std 0.905797 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81553 +trainer/policy/normal/std Std 0.628269 +trainer/policy/normal/std Max 5.54708 +trainer/policy/normal/std Min 0.338555 +trainer/policy/normal/log_std Mean 1.00092 +trainer/policy/normal/log_std Std 0.290862 +trainer/policy/normal/log_std Max 1.71327 +trainer/policy/normal/log_std Min -1.08307 +eval/num steps total 409293 +eval/num paths total 411 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.277615 +eval/Actions Std 0.866386 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66552 +time/logging (s) 0.00368584 +time/sampling batch (s) 0.279091 +time/saving (s) 0.00343395 +time/training (s) 7.94679 +time/epoch (s) 10.8985 +time/total (s) 4278.27 +Epoch -590 +---------------------------------- --------------- +2022-05-10 14:22:09.579361 PDT | [1] Epoch -589 finished +---------------------------------- --------------- +epoch -589 +replay_buffer/size 999033 +trainer/num train calls 412000 +trainer/Policy Loss -20.0179 +trainer/Log Pis Mean 24.8901 +trainer/Log Pis Std 13.2887 +trainer/Log Pis Max 62.8286 +trainer/Log Pis Min -4.49116 +trainer/policy/mean Mean -0.0137379 +trainer/policy/mean Std 0.906067 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84718 +trainer/policy/normal/std Std 0.630471 +trainer/policy/normal/std Max 5.20171 +trainer/policy/normal/std Min 0.351888 +trainer/policy/normal/log_std Mean 1.01173 +trainer/policy/normal/log_std Std 0.294512 +trainer/policy/normal/log_std Max 1.64899 +trainer/policy/normal/log_std Min -1.04444 +eval/num steps total 410293 +eval/num paths total 412 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105615 +eval/Actions Std 0.905674 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56696 +time/logging (s) 0.00389196 +time/sampling batch (s) 0.283026 +time/saving (s) 0.00397149 +time/training (s) 7.69259 +time/epoch (s) 10.5504 +time/total (s) 4288.82 +Epoch -589 +---------------------------------- --------------- +2022-05-10 14:22:20.228942 PDT | [1] Epoch -588 finished +---------------------------------- -------------- +epoch -588 +replay_buffer/size 999033 +trainer/num train calls 413000 +trainer/Policy Loss -19.5009 +trainer/Log Pis Mean 24.1848 +trainer/Log Pis Std 12.2447 +trainer/Log Pis Max 66.7002 +trainer/Log Pis Min -7.66955 +trainer/policy/mean Mean -0.03764 +trainer/policy/mean Std 0.906922 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.83639 +trainer/policy/normal/std Std 0.627301 +trainer/policy/normal/std Max 5.29127 +trainer/policy/normal/std Min 0.341824 +trainer/policy/normal/log_std Mean 1.00787 +trainer/policy/normal/log_std Std 0.294731 +trainer/policy/normal/log_std Max 1.66606 +trainer/policy/normal/log_std Min -1.07346 +eval/num steps total 411293 +eval/num paths total 413 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0818539 +eval/Actions Std 0.915428 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59286 +time/logging (s) 0.004187 +time/sampling batch (s) 0.333805 +time/saving (s) 0.0040732 +time/training (s) 7.69189 +time/epoch (s) 10.6268 +time/total (s) 4299.45 +Epoch -588 +---------------------------------- -------------- +2022-05-10 14:22:31.593075 PDT | [1] Epoch -587 finished +---------------------------------- --------------- +epoch -587 +replay_buffer/size 999033 +trainer/num train calls 414000 +trainer/Policy Loss -20.1138 +trainer/Log Pis Mean 25.3717 +trainer/Log Pis Std 13.3341 +trainer/Log Pis Max 62.2268 +trainer/Log Pis Min -8.54869 +trainer/policy/mean Mean -0.0415789 +trainer/policy/mean Std 0.905363 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87765 +trainer/policy/normal/std Std 0.66129 +trainer/policy/normal/std Max 6.89313 +trainer/policy/normal/std Min 0.314517 +trainer/policy/normal/log_std Mean 1.01975 +trainer/policy/normal/log_std Std 0.305532 +trainer/policy/normal/log_std Max 1.93052 +trainer/policy/normal/log_std Min -1.15672 +eval/num steps total 412293 +eval/num paths total 414 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0471324 +eval/Actions Std 0.912308 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72053 +time/logging (s) 0.00402403 +time/sampling batch (s) 1.08614 +time/saving (s) 0.00393405 +time/training (s) 7.52597 +time/epoch (s) 11.3406 +time/total (s) 4310.79 +Epoch -587 +---------------------------------- --------------- +2022-05-10 14:22:42.766093 PDT | [1] Epoch -586 finished +---------------------------------- --------------- +epoch -586 +replay_buffer/size 999033 +trainer/num train calls 415000 +trainer/Policy Loss -20.2254 +trainer/Log Pis Mean 24.582 +trainer/Log Pis Std 13.4048 +trainer/Log Pis Max 68.3393 +trainer/Log Pis Min -14.0848 +trainer/policy/mean Mean -0.0102283 +trainer/policy/mean Std 0.908692 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76242 +trainer/policy/normal/std Std 0.644855 +trainer/policy/normal/std Max 5.69639 +trainer/policy/normal/std Min 0.311812 +trainer/policy/normal/log_std Mean 0.976555 +trainer/policy/normal/log_std Std 0.316921 +trainer/policy/normal/log_std Max 1.73983 +trainer/policy/normal/log_std Min -1.16535 +eval/num steps total 413293 +eval/num paths total 415 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.112215 +eval/Actions Std 0.8906 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70034 +time/logging (s) 0.00388753 +time/sampling batch (s) 0.58284 +time/saving (s) 0.00367652 +time/training (s) 7.85914 +time/epoch (s) 11.1499 +time/total (s) 4321.95 +Epoch -586 +---------------------------------- --------------- +2022-05-10 14:22:53.940043 PDT | [1] Epoch -585 finished +---------------------------------- --------------- +epoch -585 +replay_buffer/size 999033 +trainer/num train calls 416000 +trainer/Policy Loss -17.4153 +trainer/Log Pis Mean 23.8637 +trainer/Log Pis Std 13.5179 +trainer/Log Pis Max 72.0881 +trainer/Log Pis Min -6.73234 +trainer/policy/mean Mean -0.0326243 +trainer/policy/mean Std 0.906181 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.78674 +trainer/policy/normal/std Std 0.601329 +trainer/policy/normal/std Max 5.40443 +trainer/policy/normal/std Min 0.374919 +trainer/policy/normal/log_std Mean 0.992604 +trainer/policy/normal/log_std Std 0.283247 +trainer/policy/normal/log_std Max 1.68722 +trainer/policy/normal/log_std Min -0.981044 +eval/num steps total 414293 +eval/num paths total 416 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.121285 +eval/Actions Std 0.894598 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69213 +time/logging (s) 0.00415192 +time/sampling batch (s) 0.832577 +time/saving (s) 0.0040825 +time/training (s) 7.61822 +time/epoch (s) 11.1512 +time/total (s) 4333.1 +Epoch -585 +---------------------------------- --------------- +2022-05-10 14:23:05.752818 PDT | [1] Epoch -584 finished +---------------------------------- --------------- +epoch -584 +replay_buffer/size 999033 +trainer/num train calls 417000 +trainer/Policy Loss -19.55 +trainer/Log Pis Mean 25.8864 +trainer/Log Pis Std 12.8573 +trainer/Log Pis Max 77.6131 +trainer/Log Pis Min -5.73138 +trainer/policy/mean Mean -0.0576624 +trainer/policy/mean Std 0.912011 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83388 +trainer/policy/normal/std Std 0.605398 +trainer/policy/normal/std Max 5.16792 +trainer/policy/normal/std Min 0.281832 +trainer/policy/normal/log_std Mean 1.01 +trainer/policy/normal/log_std Std 0.280852 +trainer/policy/normal/log_std Max 1.64247 +trainer/policy/normal/log_std Min -1.26645 +eval/num steps total 415293 +eval/num paths total 417 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.114906 +eval/Actions Std 0.901158 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69803 +time/logging (s) 0.0038955 +time/sampling batch (s) 1.33657 +time/saving (s) 0.00398789 +time/training (s) 7.74658 +time/epoch (s) 11.7891 +time/total (s) 4344.89 +Epoch -584 +---------------------------------- --------------- +2022-05-10 14:23:16.796896 PDT | [1] Epoch -583 finished +---------------------------------- --------------- +epoch -583 +replay_buffer/size 999033 +trainer/num train calls 418000 +trainer/Policy Loss -19.8555 +trainer/Log Pis Mean 25.0445 +trainer/Log Pis Std 13.4458 +trainer/Log Pis Max 70.5254 +trainer/Log Pis Min -14.1997 +trainer/policy/mean Mean -0.0408069 +trainer/policy/mean Std 0.907298 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86741 +trainer/policy/normal/std Std 0.631886 +trainer/policy/normal/std Max 6.46012 +trainer/policy/normal/std Min 0.226656 +trainer/policy/normal/log_std Mean 1.01811 +trainer/policy/normal/log_std Std 0.300978 +trainer/policy/normal/log_std Max 1.86565 +trainer/policy/normal/log_std Min -1.48432 +eval/num steps total 416293 +eval/num paths total 418 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.15467 +eval/Actions Std 0.926912 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83665 +time/logging (s) 0.00375146 +time/sampling batch (s) 0.834148 +time/saving (s) 0.00366842 +time/training (s) 7.34256 +time/epoch (s) 11.0208 +time/total (s) 4355.92 +Epoch -583 +---------------------------------- --------------- +2022-05-10 14:23:27.093646 PDT | [1] Epoch -582 finished +---------------------------------- --------------- +epoch -582 +replay_buffer/size 999033 +trainer/num train calls 419000 +trainer/Policy Loss -19.7874 +trainer/Log Pis Mean 24.5859 +trainer/Log Pis Std 13.3312 +trainer/Log Pis Max 70.3838 +trainer/Log Pis Min -10.3701 +trainer/policy/mean Mean -0.0165507 +trainer/policy/mean Std 0.909998 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82529 +trainer/policy/normal/std Std 0.603747 +trainer/policy/normal/std Max 4.96267 +trainer/policy/normal/std Min 0.347414 +trainer/policy/normal/log_std Mean 1.00618 +trainer/policy/normal/log_std Std 0.285737 +trainer/policy/normal/log_std Max 1.60194 +trainer/policy/normal/log_std Min -1.05724 +eval/num steps total 417293 +eval/num paths total 419 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0922496 +eval/Actions Std 0.917057 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.9726 +time/logging (s) 0.00378356 +time/sampling batch (s) 0.292906 +time/saving (s) 0.00343895 +time/training (s) 7.00177 +time/epoch (s) 10.2745 +time/total (s) 4366.19 +Epoch -582 +---------------------------------- --------------- +2022-05-10 14:23:37.670240 PDT | [1] Epoch -581 finished +---------------------------------- --------------- +epoch -581 +replay_buffer/size 999033 +trainer/num train calls 420000 +trainer/Policy Loss -19.8762 +trainer/Log Pis Mean 23.3251 +trainer/Log Pis Std 13.4217 +trainer/Log Pis Max 65.9681 +trainer/Log Pis Min -12.3843 +trainer/policy/mean Mean -0.0484344 +trainer/policy/mean Std 0.905129 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78258 +trainer/policy/normal/std Std 0.610903 +trainer/policy/normal/std Max 6.44103 +trainer/policy/normal/std Min 0.371338 +trainer/policy/normal/log_std Mean 0.990355 +trainer/policy/normal/log_std Std 0.286013 +trainer/policy/normal/log_std Max 1.86269 +trainer/policy/normal/log_std Min -0.990643 +eval/num steps total 418293 +eval/num paths total 420 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0362267 +eval/Actions Std 0.903553 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6331 +time/logging (s) 0.00409814 +time/sampling batch (s) 0.282981 +time/saving (s) 0.00388096 +time/training (s) 7.63055 +time/epoch (s) 10.5546 +time/total (s) 4376.75 +Epoch -581 +---------------------------------- --------------- +2022-05-10 14:23:48.386852 PDT | [1] Epoch -580 finished +---------------------------------- --------------- +epoch -580 +replay_buffer/size 999033 +trainer/num train calls 421000 +trainer/Policy Loss -20.2194 +trainer/Log Pis Mean 24.9911 +trainer/Log Pis Std 13.3994 +trainer/Log Pis Max 68.5542 +trainer/Log Pis Min -8.69359 +trainer/policy/mean Mean -0.0358323 +trainer/policy/mean Std 0.904957 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82826 +trainer/policy/normal/std Std 0.663562 +trainer/policy/normal/std Max 6.65516 +trainer/policy/normal/std Min 0.254935 +trainer/policy/normal/log_std Mean 0.999376 +trainer/policy/normal/log_std Std 0.322227 +trainer/policy/normal/log_std Max 1.89539 +trainer/policy/normal/log_std Min -1.36675 +eval/num steps total 419293 +eval/num paths total 421 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.10072 +eval/Actions Std 0.902704 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50629 +time/logging (s) 0.00400885 +time/sampling batch (s) 0.529283 +time/saving (s) 0.00371443 +time/training (s) 7.65059 +time/epoch (s) 10.6939 +time/total (s) 4387.45 +Epoch -580 +---------------------------------- --------------- +2022-05-10 14:23:58.563830 PDT | [1] Epoch -579 finished +---------------------------------- --------------- +epoch -579 +replay_buffer/size 999033 +trainer/num train calls 422000 +trainer/Policy Loss -19.6339 +trainer/Log Pis Mean 25.5633 +trainer/Log Pis Std 13.8323 +trainer/Log Pis Max 67.0878 +trainer/Log Pis Min -4.35769 +trainer/policy/mean Mean -0.0276765 +trainer/policy/mean Std 0.902464 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85548 +trainer/policy/normal/std Std 0.653276 +trainer/policy/normal/std Max 7.18215 +trainer/policy/normal/std Min 0.33987 +trainer/policy/normal/log_std Mean 1.01274 +trainer/policy/normal/log_std Std 0.302559 +trainer/policy/normal/log_std Max 1.9716 +trainer/policy/normal/log_std Min -1.07919 +eval/num steps total 420293 +eval/num paths total 422 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0236544 +eval/Actions Std 0.903735 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71464 +time/logging (s) 0.00372115 +time/sampling batch (s) 0.532364 +time/saving (s) 0.00344697 +time/training (s) 6.90025 +time/epoch (s) 10.1544 +time/total (s) 4397.61 +Epoch -579 +---------------------------------- --------------- +2022-05-10 14:24:08.679437 PDT | [1] Epoch -578 finished +---------------------------------- --------------- +epoch -578 +replay_buffer/size 999033 +trainer/num train calls 423000 +trainer/Policy Loss -20.4867 +trainer/Log Pis Mean 23.6072 +trainer/Log Pis Std 13.4013 +trainer/Log Pis Max 74.3857 +trainer/Log Pis Min -9.87381 +trainer/policy/mean Mean -0.042675 +trainer/policy/mean Std 0.90545 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78357 +trainer/policy/normal/std Std 0.621012 +trainer/policy/normal/std Max 4.8701 +trainer/policy/normal/std Min 0.26998 +trainer/policy/normal/log_std Mean 0.988003 +trainer/policy/normal/log_std Std 0.300693 +trainer/policy/normal/log_std Max 1.58311 +trainer/policy/normal/log_std Min -1.30941 +eval/num steps total 421293 +eval/num paths total 423 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.344169 +eval/Actions Std 0.837436 +eval/Actions Max 0.999996 +eval/Actions Min -0.999961 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66627 +time/logging (s) 0.00371998 +time/sampling batch (s) 0.279569 +time/saving (s) 0.00346745 +time/training (s) 7.14059 +time/epoch (s) 10.0936 +time/total (s) 4407.7 +Epoch -578 +---------------------------------- --------------- +2022-05-10 14:24:18.884843 PDT | [1] Epoch -577 finished +---------------------------------- --------------- +epoch -577 +replay_buffer/size 999033 +trainer/num train calls 424000 +trainer/Policy Loss -20.7103 +trainer/Log Pis Mean 24.9331 +trainer/Log Pis Std 14.2642 +trainer/Log Pis Max 76.4469 +trainer/Log Pis Min -6.35783 +trainer/policy/mean Mean -0.0380603 +trainer/policy/mean Std 0.905675 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8185 +trainer/policy/normal/std Std 0.630873 +trainer/policy/normal/std Max 5.7127 +trainer/policy/normal/std Min 0.347387 +trainer/policy/normal/log_std Mean 0.999863 +trainer/policy/normal/log_std Std 0.304793 +trainer/policy/normal/log_std Max 1.74269 +trainer/policy/normal/log_std Min -1.05732 +eval/num steps total 422293 +eval/num paths total 424 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00862537 +eval/Actions Std 0.902032 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58675 +time/logging (s) 0.00380879 +time/sampling batch (s) 0.279883 +time/saving (s) 0.00357414 +time/training (s) 7.30954 +time/epoch (s) 10.1836 +time/total (s) 4417.89 +Epoch -577 +---------------------------------- --------------- +2022-05-10 14:24:28.692310 PDT | [1] Epoch -576 finished +---------------------------------- --------------- +epoch -576 +replay_buffer/size 999033 +trainer/num train calls 425000 +trainer/Policy Loss -19.9347 +trainer/Log Pis Mean 22.9173 +trainer/Log Pis Std 13.2401 +trainer/Log Pis Max 81.0656 +trainer/Log Pis Min -11.2084 +trainer/policy/mean Mean -0.0542087 +trainer/policy/mean Std 0.901289 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83917 +trainer/policy/normal/std Std 0.631661 +trainer/policy/normal/std Max 6.36413 +trainer/policy/normal/std Min 0.226844 +trainer/policy/normal/log_std Mean 1.00857 +trainer/policy/normal/log_std Std 0.297751 +trainer/policy/normal/log_std Max 1.85068 +trainer/policy/normal/log_std Min -1.48349 +eval/num steps total 423293 +eval/num paths total 425 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103867 +eval/Actions Std 0.910709 +eval/Actions Max 0.999996 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74086 +time/logging (s) 0.00416702 +time/sampling batch (s) 0.280001 +time/saving (s) 0.00410947 +time/training (s) 6.75634 +time/epoch (s) 9.78548 +time/total (s) 4427.68 +Epoch -576 +---------------------------------- --------------- +2022-05-10 14:24:38.892185 PDT | [1] Epoch -575 finished +---------------------------------- --------------- +epoch -575 +replay_buffer/size 999033 +trainer/num train calls 426000 +trainer/Policy Loss -20.6192 +trainer/Log Pis Mean 24.8515 +trainer/Log Pis Std 12.9847 +trainer/Log Pis Max 69.2038 +trainer/Log Pis Min -9.49782 +trainer/policy/mean Mean -0.0501236 +trainer/policy/mean Std 0.9045 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81994 +trainer/policy/normal/std Std 0.63343 +trainer/policy/normal/std Max 6.74134 +trainer/policy/normal/std Min 0.314794 +trainer/policy/normal/log_std Mean 1.00163 +trainer/policy/normal/log_std Std 0.29635 +trainer/policy/normal/log_std Max 1.90826 +trainer/policy/normal/log_std Min -1.15584 +eval/num steps total 424293 +eval/num paths total 426 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.159538 +eval/Actions Std 0.901488 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3547 +time/logging (s) 0.00375835 +time/sampling batch (s) 0.529918 +time/saving (s) 0.0034458 +time/training (s) 7.28525 +time/epoch (s) 10.1771 +time/total (s) 4437.86 +Epoch -575 +---------------------------------- --------------- +2022-05-10 14:24:48.988362 PDT | [1] Epoch -574 finished +---------------------------------- --------------- +epoch -574 +replay_buffer/size 999033 +trainer/num train calls 427000 +trainer/Policy Loss -19.3717 +trainer/Log Pis Mean 24.3723 +trainer/Log Pis Std 13.3241 +trainer/Log Pis Max 68.6251 +trainer/Log Pis Min -7.54804 +trainer/policy/mean Mean -0.0371748 +trainer/policy/mean Std 0.907261 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77627 +trainer/policy/normal/std Std 0.618586 +trainer/policy/normal/std Max 6.08518 +trainer/policy/normal/std Min 0.346636 +trainer/policy/normal/log_std Mean 0.985553 +trainer/policy/normal/log_std Std 0.299677 +trainer/policy/normal/log_std Max 1.80586 +trainer/policy/normal/log_std Min -1.05948 +eval/num steps total 425293 +eval/num paths total 427 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0195035 +eval/Actions Std 0.929029 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74987 +time/logging (s) 0.00372468 +time/sampling batch (s) 0.278602 +time/saving (s) 0.00343391 +time/training (s) 7.03841 +time/epoch (s) 10.074 +time/total (s) 4447.93 +Epoch -574 +---------------------------------- --------------- +2022-05-10 14:24:59.759211 PDT | [1] Epoch -573 finished +---------------------------------- --------------- +epoch -573 +replay_buffer/size 999033 +trainer/num train calls 428000 +trainer/Policy Loss -20.4831 +trainer/Log Pis Mean 23.2437 +trainer/Log Pis Std 12.546 +trainer/Log Pis Max 68.0773 +trainer/Log Pis Min -5.57672 +trainer/policy/mean Mean -0.0408233 +trainer/policy/mean Std 0.903634 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.73276 +trainer/policy/normal/std Std 0.618819 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.267569 +trainer/policy/normal/log_std Mean 0.968854 +trainer/policy/normal/log_std Std 0.304233 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.31838 +eval/num steps total 426293 +eval/num paths total 428 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.174366 +eval/Actions Std 0.884263 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69683 +time/logging (s) 0.00412283 +time/sampling batch (s) 0.279118 +time/saving (s) 0.0038484 +time/training (s) 7.76504 +time/epoch (s) 10.749 +time/total (s) 4458.69 +Epoch -573 +---------------------------------- --------------- +2022-05-10 14:25:09.989597 PDT | [1] Epoch -572 finished +---------------------------------- --------------- +epoch -572 +replay_buffer/size 999033 +trainer/num train calls 429000 +trainer/Policy Loss -19.4459 +trainer/Log Pis Mean 24.7841 +trainer/Log Pis Std 13.6559 +trainer/Log Pis Max 83.2799 +trainer/Log Pis Min -8.66872 +trainer/policy/mean Mean -0.0369364 +trainer/policy/mean Std 0.905442 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86656 +trainer/policy/normal/std Std 0.662787 +trainer/policy/normal/std Max 6.02381 +trainer/policy/normal/std Min 0.339845 +trainer/policy/normal/log_std Mean 1.01414 +trainer/policy/normal/log_std Std 0.316269 +trainer/policy/normal/log_std Max 1.79572 +trainer/policy/normal/log_std Min -1.07927 +eval/num steps total 427293 +eval/num paths total 429 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0100326 +eval/Actions Std 0.863305 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67621 +time/logging (s) 0.00387899 +time/sampling batch (s) 0.788463 +time/saving (s) 0.00396521 +time/training (s) 6.73506 +time/epoch (s) 10.2076 +time/total (s) 4468.9 +Epoch -572 +---------------------------------- --------------- +2022-05-10 14:25:20.008750 PDT | [1] Epoch -571 finished +---------------------------------- --------------- +epoch -571 +replay_buffer/size 999033 +trainer/num train calls 430000 +trainer/Policy Loss -19.5968 +trainer/Log Pis Mean 24.2139 +trainer/Log Pis Std 13.0971 +trainer/Log Pis Max 64.9313 +trainer/Log Pis Min -8.62029 +trainer/policy/mean Mean -0.0401128 +trainer/policy/mean Std 0.906206 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81261 +trainer/policy/normal/std Std 0.63457 +trainer/policy/normal/std Max 5.12013 +trainer/policy/normal/std Min 0.281995 +trainer/policy/normal/log_std Mean 0.995984 +trainer/policy/normal/log_std Std 0.314798 +trainer/policy/normal/log_std Max 1.63318 +trainer/policy/normal/log_std Min -1.26587 +eval/num steps total 428293 +eval/num paths total 430 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116262 +eval/Actions Std 0.917625 +eval/Actions Max 0.999974 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 3.01303 +time/logging (s) 0.00370771 +time/sampling batch (s) 0.282365 +time/saving (s) 0.00341309 +time/training (s) 6.6939 +time/epoch (s) 9.99642 +time/total (s) 4478.9 +Epoch -571 +---------------------------------- --------------- +2022-05-10 14:25:30.659295 PDT | [1] Epoch -570 finished +---------------------------------- --------------- +epoch -570 +replay_buffer/size 999033 +trainer/num train calls 431000 +trainer/Policy Loss -17.8018 +trainer/Log Pis Mean 23.0217 +trainer/Log Pis Std 13.1477 +trainer/Log Pis Max 68.2908 +trainer/Log Pis Min -7.68901 +trainer/policy/mean Mean -0.0402455 +trainer/policy/mean Std 0.898243 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.78599 +trainer/policy/normal/std Std 0.647631 +trainer/policy/normal/std Max 5.33543 +trainer/policy/normal/std Min 0.213982 +trainer/policy/normal/log_std Mean 0.984263 +trainer/policy/normal/log_std Std 0.324431 +trainer/policy/normal/log_std Max 1.67437 +trainer/policy/normal/log_std Min -1.54186 +eval/num steps total 429293 +eval/num paths total 431 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.345572 +eval/Actions Std 0.866554 +eval/Actions Max 0.999986 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57531 +time/logging (s) 0.00372917 +time/sampling batch (s) 0.278683 +time/saving (s) 0.00343773 +time/training (s) 7.76712 +time/epoch (s) 10.6283 +time/total (s) 4489.53 +Epoch -570 +---------------------------------- --------------- +2022-05-10 14:25:40.381903 PDT | [1] Epoch -569 finished +---------------------------------- --------------- +epoch -569 +replay_buffer/size 999033 +trainer/num train calls 432000 +trainer/Policy Loss -19.6584 +trainer/Log Pis Mean 24.4769 +trainer/Log Pis Std 13.3735 +trainer/Log Pis Max 73.9242 +trainer/Log Pis Min -9.82618 +trainer/policy/mean Mean -0.0485079 +trainer/policy/mean Std 0.908266 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85113 +trainer/policy/normal/std Std 0.632057 +trainer/policy/normal/std Max 5.49397 +trainer/policy/normal/std Min 0.305152 +trainer/policy/normal/log_std Mean 1.01297 +trainer/policy/normal/log_std Std 0.295967 +trainer/policy/normal/log_std Max 1.70365 +trainer/policy/normal/log_std Min -1.18695 +eval/num steps total 430293 +eval/num paths total 432 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.113955 +eval/Actions Std 0.908231 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44733 +time/logging (s) 0.00416572 +time/sampling batch (s) 0.278857 +time/saving (s) 0.00380779 +time/training (s) 6.9666 +time/epoch (s) 9.70077 +time/total (s) 4499.23 +Epoch -569 +---------------------------------- --------------- +2022-05-10 14:25:51.376152 PDT | [1] Epoch -568 finished +---------------------------------- --------------- +epoch -568 +replay_buffer/size 999033 +trainer/num train calls 433000 +trainer/Policy Loss -19.6133 +trainer/Log Pis Mean 25.4377 +trainer/Log Pis Std 13.6755 +trainer/Log Pis Max 71.8235 +trainer/Log Pis Min -9.81365 +trainer/policy/mean Mean -0.0637796 +trainer/policy/mean Std 0.908154 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8608 +trainer/policy/normal/std Std 0.656183 +trainer/policy/normal/std Max 6.28924 +trainer/policy/normal/std Min 0.312338 +trainer/policy/normal/log_std Mean 1.01289 +trainer/policy/normal/log_std Std 0.312715 +trainer/policy/normal/log_std Max 1.83884 +trainer/policy/normal/log_std Min -1.16367 +eval/num steps total 431293 +eval/num paths total 433 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.165867 +eval/Actions Std 0.862622 +eval/Actions Max 0.99999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36314 +time/logging (s) 0.00391894 +time/sampling batch (s) 0.277824 +time/saving (s) 0.00379318 +time/training (s) 8.32265 +time/epoch (s) 10.9713 +time/total (s) 4510.21 +Epoch -568 +---------------------------------- --------------- +2022-05-10 14:26:03.107015 PDT | [1] Epoch -567 finished +---------------------------------- --------------- +epoch -567 +replay_buffer/size 999033 +trainer/num train calls 434000 +trainer/Policy Loss -19.4683 +trainer/Log Pis Mean 24.0187 +trainer/Log Pis Std 13.5374 +trainer/Log Pis Max 70.4025 +trainer/Log Pis Min -5.53133 +trainer/policy/mean Mean -0.0303645 +trainer/policy/mean Std 0.908924 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88558 +trainer/policy/normal/std Std 0.633565 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.384368 +trainer/policy/normal/log_std Mean 1.02641 +trainer/policy/normal/log_std Std 0.287263 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.956155 +eval/num steps total 432293 +eval/num paths total 434 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.168226 +eval/Actions Std 0.817647 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70349 +time/logging (s) 0.0036935 +time/sampling batch (s) 0.277806 +time/saving (s) 0.00342814 +time/training (s) 8.72012 +time/epoch (s) 11.7085 +time/total (s) 4521.92 +Epoch -567 +---------------------------------- --------------- +2022-05-10 14:26:13.268874 PDT | [1] Epoch -566 finished +---------------------------------- --------------- +epoch -566 +replay_buffer/size 999033 +trainer/num train calls 435000 +trainer/Policy Loss -17.9958 +trainer/Log Pis Mean 24.554 +trainer/Log Pis Std 13.2752 +trainer/Log Pis Max 75.8657 +trainer/Log Pis Min -9.48187 +trainer/policy/mean Mean -0.026776 +trainer/policy/mean Std 0.904226 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82918 +trainer/policy/normal/std Std 0.64317 +trainer/policy/normal/std Max 6.41495 +trainer/policy/normal/std Min 0.321165 +trainer/policy/normal/log_std Mean 1.00261 +trainer/policy/normal/log_std Std 0.308975 +trainer/policy/normal/log_std Max 1.85863 +trainer/policy/normal/log_std Min -1.1358 +eval/num steps total 433293 +eval/num paths total 435 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138209 +eval/Actions Std 0.898209 +eval/Actions Max 0.999996 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45503 +time/logging (s) 0.00378512 +time/sampling batch (s) 0.278392 +time/saving (s) 0.00345381 +time/training (s) 7.39925 +time/epoch (s) 10.1399 +time/total (s) 4532.06 +Epoch -566 +---------------------------------- --------------- +2022-05-10 14:26:23.166888 PDT | [1] Epoch -565 finished +---------------------------------- --------------- +epoch -565 +replay_buffer/size 999033 +trainer/num train calls 436000 +trainer/Policy Loss -18.831 +trainer/Log Pis Mean 23.6892 +trainer/Log Pis Std 13.2902 +trainer/Log Pis Max 70.4049 +trainer/Log Pis Min -10.0415 +trainer/policy/mean Mean -0.0445278 +trainer/policy/mean Std 0.905495 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82693 +trainer/policy/normal/std Std 0.619431 +trainer/policy/normal/std Max 5.60775 +trainer/policy/normal/std Min 0.206233 +trainer/policy/normal/log_std Mean 1.00423 +trainer/policy/normal/log_std Std 0.299931 +trainer/policy/normal/log_std Max 1.72415 +trainer/policy/normal/log_std Min -1.57875 +eval/num steps total 434293 +eval/num paths total 436 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.299425 +eval/Actions Std 0.88027 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72948 +time/logging (s) 0.00415813 +time/sampling batch (s) 0.279535 +time/saving (s) 0.00398859 +time/training (s) 6.85898 +time/epoch (s) 9.87614 +time/total (s) 4541.94 +Epoch -565 +---------------------------------- --------------- +2022-05-10 14:26:32.963825 PDT | [1] Epoch -564 finished +---------------------------------- --------------- +epoch -564 +replay_buffer/size 999033 +trainer/num train calls 437000 +trainer/Policy Loss -19.4085 +trainer/Log Pis Mean 25.1236 +trainer/Log Pis Std 13.2233 +trainer/Log Pis Max 74.5536 +trainer/Log Pis Min -4.50075 +trainer/policy/mean Mean -0.00288126 +trainer/policy/mean Std 0.905388 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81158 +trainer/policy/normal/std Std 0.655908 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.269085 +trainer/policy/normal/log_std Mean 0.994644 +trainer/policy/normal/log_std Std 0.315358 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.31273 +eval/num steps total 435293 +eval/num paths total 437 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.52233 +eval/Actions Std 0.743759 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57377 +time/logging (s) 0.00377887 +time/sampling batch (s) 0.280297 +time/saving (s) 0.00359356 +time/training (s) 6.91253 +time/epoch (s) 9.77397 +time/total (s) 4551.72 +Epoch -564 +---------------------------------- --------------- +2022-05-10 14:26:42.141682 PDT | [1] Epoch -563 finished +---------------------------------- --------------- +epoch -563 +replay_buffer/size 999033 +trainer/num train calls 438000 +trainer/Policy Loss -20.4764 +trainer/Log Pis Mean 23.0823 +trainer/Log Pis Std 13.5219 +trainer/Log Pis Max 69.2392 +trainer/Log Pis Min -9.02752 +trainer/policy/mean Mean -0.073456 +trainer/policy/mean Std 0.902867 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77913 +trainer/policy/normal/std Std 0.625695 +trainer/policy/normal/std Max 5.3256 +trainer/policy/normal/std Min 0.339385 +trainer/policy/normal/log_std Mean 0.985281 +trainer/policy/normal/log_std Std 0.306614 +trainer/policy/normal/log_std Max 1.67253 +trainer/policy/normal/log_std Min -1.08062 +eval/num steps total 436293 +eval/num paths total 438 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.28154 +eval/Actions Std 0.887332 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70391 +time/logging (s) 0.00373748 +time/sampling batch (s) 0.279473 +time/saving (s) 0.00343689 +time/training (s) 6.16508 +time/epoch (s) 9.15564 +time/total (s) 4560.88 +Epoch -563 +---------------------------------- --------------- +2022-05-10 14:26:53.232440 PDT | [1] Epoch -562 finished +---------------------------------- --------------- +epoch -562 +replay_buffer/size 999033 +trainer/num train calls 439000 +trainer/Policy Loss -20.5853 +trainer/Log Pis Mean 24.1898 +trainer/Log Pis Std 13.6859 +trainer/Log Pis Max 79.1994 +trainer/Log Pis Min -7.80375 +trainer/policy/mean Mean -0.0280338 +trainer/policy/mean Std 0.906121 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84247 +trainer/policy/normal/std Std 0.63747 +trainer/policy/normal/std Max 5.44558 +trainer/policy/normal/std Min 0.296503 +trainer/policy/normal/log_std Mean 1.00857 +trainer/policy/normal/log_std Std 0.303094 +trainer/policy/normal/log_std Max 1.6948 +trainer/policy/normal/log_std Min -1.2157 +eval/num steps total 437293 +eval/num paths total 439 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0445849 +eval/Actions Std 0.896017 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44213 +time/logging (s) 0.0037862 +time/sampling batch (s) 0.279401 +time/saving (s) 0.00337709 +time/training (s) 8.34017 +time/epoch (s) 11.0689 +time/total (s) 4571.95 +Epoch -562 +---------------------------------- --------------- +2022-05-10 14:27:03.368890 PDT | [1] Epoch -561 finished +---------------------------------- --------------- +epoch -561 +replay_buffer/size 999033 +trainer/num train calls 440000 +trainer/Policy Loss -19.5607 +trainer/Log Pis Mean 24.3864 +trainer/Log Pis Std 13.5918 +trainer/Log Pis Max 73.5154 +trainer/Log Pis Min -9.98811 +trainer/policy/mean Mean -0.0230409 +trainer/policy/mean Std 0.909668 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.76806 +trainer/policy/normal/std Std 0.643934 +trainer/policy/normal/std Max 4.96994 +trainer/policy/normal/std Min 0.281758 +trainer/policy/normal/log_std Mean 0.977901 +trainer/policy/normal/log_std Std 0.321331 +trainer/policy/normal/log_std Max 1.60341 +trainer/policy/normal/log_std Min -1.26671 +eval/num steps total 438293 +eval/num paths total 440 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.347797 +eval/Actions Std 0.546419 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81394 +time/logging (s) 0.00393876 +time/sampling batch (s) 0.275651 +time/saving (s) 0.00356057 +time/training (s) 7.01771 +time/epoch (s) 10.1148 +time/total (s) 4582.07 +Epoch -561 +---------------------------------- --------------- +2022-05-10 14:27:13.501588 PDT | [1] Epoch -560 finished +---------------------------------- --------------- +epoch -560 +replay_buffer/size 999033 +trainer/num train calls 441000 +trainer/Policy Loss -20.5352 +trainer/Log Pis Mean 23.982 +trainer/Log Pis Std 12.5938 +trainer/Log Pis Max 65.3781 +trainer/Log Pis Min -7.22543 +trainer/policy/mean Mean -0.0185795 +trainer/policy/mean Std 0.903588 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80282 +trainer/policy/normal/std Std 0.626262 +trainer/policy/normal/std Max 5.09108 +trainer/policy/normal/std Min 0.278127 +trainer/policy/normal/log_std Mean 0.995449 +trainer/policy/normal/log_std Std 0.297257 +trainer/policy/normal/log_std Max 1.62749 +trainer/policy/normal/log_std Min -1.27968 +eval/num steps total 439293 +eval/num paths total 441 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.214826 +eval/Actions Std 0.811018 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7712 +time/logging (s) 0.00413958 +time/sampling batch (s) 0.277654 +time/saving (s) 0.00400374 +time/training (s) 7.05369 +time/epoch (s) 10.1107 +time/total (s) 4592.18 +Epoch -560 +---------------------------------- --------------- +2022-05-10 14:27:24.038308 PDT | [1] Epoch -559 finished +---------------------------------- --------------- +epoch -559 +replay_buffer/size 999033 +trainer/num train calls 442000 +trainer/Policy Loss -18.9726 +trainer/Log Pis Mean 24.19 +trainer/Log Pis Std 13.0508 +trainer/Log Pis Max 59.7931 +trainer/Log Pis Min -8.84501 +trainer/policy/mean Mean -0.0369287 +trainer/policy/mean Std 0.908097 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81262 +trainer/policy/normal/std Std 0.659262 +trainer/policy/normal/std Max 6.00464 +trainer/policy/normal/std Min 0.266885 +trainer/policy/normal/log_std Mean 0.99371 +trainer/policy/normal/log_std Std 0.322725 +trainer/policy/normal/log_std Max 1.79253 +trainer/policy/normal/log_std Min -1.32094 +eval/num steps total 440293 +eval/num paths total 442 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117374 +eval/Actions Std 0.905771 +eval/Actions Max 0.99998 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70562 +time/logging (s) 0.00369527 +time/sampling batch (s) 0.529256 +time/saving (s) 0.00361316 +time/training (s) 7.27135 +time/epoch (s) 10.5135 +time/total (s) 4602.7 +Epoch -559 +---------------------------------- --------------- +2022-05-10 14:27:33.582876 PDT | [1] Epoch -558 finished +---------------------------------- --------------- +epoch -558 +replay_buffer/size 999033 +trainer/num train calls 443000 +trainer/Policy Loss -20.0655 +trainer/Log Pis Mean 25.6054 +trainer/Log Pis Std 13.5485 +trainer/Log Pis Max 67.4685 +trainer/Log Pis Min -7.38548 +trainer/policy/mean Mean -0.0526253 +trainer/policy/mean Std 0.911128 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79205 +trainer/policy/normal/std Std 0.629448 +trainer/policy/normal/std Max 5.82612 +trainer/policy/normal/std Min 0.282981 +trainer/policy/normal/log_std Mean 0.990152 +trainer/policy/normal/log_std Std 0.304955 +trainer/policy/normal/log_std Max 1.76235 +trainer/policy/normal/log_std Min -1.26238 +eval/num steps total 441293 +eval/num paths total 443 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0356288 +eval/Actions Std 0.915517 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51541 +time/logging (s) 0.00366739 +time/sampling batch (s) 0.279012 +time/saving (s) 0.00341776 +time/training (s) 6.72072 +time/epoch (s) 9.52223 +time/total (s) 4612.22 +Epoch -558 +---------------------------------- --------------- +2022-05-10 14:27:43.744598 PDT | [1] Epoch -557 finished +---------------------------------- ---------------- +epoch -557 +replay_buffer/size 999033 +trainer/num train calls 444000 +trainer/Policy Loss -19.7001 +trainer/Log Pis Mean 25.8785 +trainer/Log Pis Std 13.5329 +trainer/Log Pis Max 83.5671 +trainer/Log Pis Min -11.7022 +trainer/policy/mean Mean -0.06216 +trainer/policy/mean Std 0.90884 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88635 +trainer/policy/normal/std Std 0.624761 +trainer/policy/normal/std Max 5.94548 +trainer/policy/normal/std Min 0.282079 +trainer/policy/normal/log_std Mean 1.02617 +trainer/policy/normal/log_std Std 0.294479 +trainer/policy/normal/log_std Max 1.78263 +trainer/policy/normal/log_std Min -1.26557 +eval/num steps total 442293 +eval/num paths total 444 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000525047 +eval/Actions Std 0.89741 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71643 +time/logging (s) 0.00374758 +time/sampling batch (s) 0.279474 +time/saving (s) 0.0034201 +time/training (s) 7.13656 +time/epoch (s) 10.1396 +time/total (s) 4622.36 +Epoch -557 +---------------------------------- ---------------- +2022-05-10 14:27:53.646051 PDT | [1] Epoch -556 finished +---------------------------------- --------------- +epoch -556 +replay_buffer/size 999033 +trainer/num train calls 445000 +trainer/Policy Loss -18.9504 +trainer/Log Pis Mean 24.2676 +trainer/Log Pis Std 13.3511 +trainer/Log Pis Max 70.3556 +trainer/Log Pis Min -14.582 +trainer/policy/mean Mean -0.0461826 +trainer/policy/mean Std 0.906473 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.86075 +trainer/policy/normal/std Std 0.65095 +trainer/policy/normal/std Max 5.75573 +trainer/policy/normal/std Min 0.340976 +trainer/policy/normal/log_std Mean 1.01266 +trainer/policy/normal/log_std Std 0.314869 +trainer/policy/normal/log_std Max 1.7502 +trainer/policy/normal/log_std Min -1.07594 +eval/num steps total 443293 +eval/num paths total 445 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.487047 +eval/Actions Std 0.717897 +eval/Actions Max 1 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48673 +time/logging (s) 0.00375689 +time/sampling batch (s) 0.279482 +time/saving (s) 0.0034384 +time/training (s) 7.10582 +time/epoch (s) 9.87923 +time/total (s) 4632.25 +Epoch -556 +---------------------------------- --------------- +2022-05-10 14:28:03.717739 PDT | [1] Epoch -555 finished +---------------------------------- --------------- +epoch -555 +replay_buffer/size 999033 +trainer/num train calls 446000 +trainer/Policy Loss -20.5718 +trainer/Log Pis Mean 24.899 +trainer/Log Pis Std 13.4342 +trainer/Log Pis Max 73.6789 +trainer/Log Pis Min -3.5488 +trainer/policy/mean Mean -0.0420301 +trainer/policy/mean Std 0.911115 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82552 +trainer/policy/normal/std Std 0.626331 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.271289 +trainer/policy/normal/log_std Mean 1.00545 +trainer/policy/normal/log_std Std 0.286742 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.30457 +eval/num steps total 444293 +eval/num paths total 446 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.169734 +eval/Actions Std 0.894879 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6267 +time/logging (s) 0.00423907 +time/sampling batch (s) 0.281579 +time/saving (s) 0.00381178 +time/training (s) 7.13341 +time/epoch (s) 10.0497 +time/total (s) 4642.3 +Epoch -555 +---------------------------------- --------------- +2022-05-10 14:28:14.439519 PDT | [1] Epoch -554 finished +---------------------------------- --------------- +epoch -554 +replay_buffer/size 999033 +trainer/num train calls 447000 +trainer/Policy Loss -19.3664 +trainer/Log Pis Mean 24.4435 +trainer/Log Pis Std 13.3749 +trainer/Log Pis Max 69.3251 +trainer/Log Pis Min -6.62498 +trainer/policy/mean Mean -0.017531 +trainer/policy/mean Std 0.908517 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88195 +trainer/policy/normal/std Std 0.660075 +trainer/policy/normal/std Max 6.93147 +trainer/policy/normal/std Min 0.27303 +trainer/policy/normal/log_std Mean 1.02266 +trainer/policy/normal/log_std Std 0.297839 +trainer/policy/normal/log_std Max 1.93607 +trainer/policy/normal/log_std Min -1.29818 +eval/num steps total 445293 +eval/num paths total 447 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.162762 +eval/Actions Std 0.900195 +eval/Actions Max 0.999979 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61891 +time/logging (s) 0.00389904 +time/sampling batch (s) 0.279714 +time/saving (s) 0.00367231 +time/training (s) 7.79257 +time/epoch (s) 10.6988 +time/total (s) 4653 +Epoch -554 +---------------------------------- --------------- +2022-05-10 14:28:25.237350 PDT | [1] Epoch -553 finished +---------------------------------- --------------- +epoch -553 +replay_buffer/size 999033 +trainer/num train calls 448000 +trainer/Policy Loss -19.4831 +trainer/Log Pis Mean 25.523 +trainer/Log Pis Std 13.1205 +trainer/Log Pis Max 68.0194 +trainer/Log Pis Min -8.72462 +trainer/policy/mean Mean -0.0309466 +trainer/policy/mean Std 0.911691 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79705 +trainer/policy/normal/std Std 0.624192 +trainer/policy/normal/std Max 6.43005 +trainer/policy/normal/std Min 0.262368 +trainer/policy/normal/log_std Mean 0.993171 +trainer/policy/normal/log_std Std 0.299283 +trainer/policy/normal/log_std Max 1.86098 +trainer/policy/normal/log_std Min -1.33801 +eval/num steps total 446293 +eval/num paths total 448 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123911 +eval/Actions Std 0.88711 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55248 +time/logging (s) 0.00370298 +time/sampling batch (s) 0.277471 +time/saving (s) 0.0034272 +time/training (s) 7.93825 +time/epoch (s) 10.7753 +time/total (s) 4663.78 +Epoch -553 +---------------------------------- --------------- +2022-05-10 14:28:35.308840 PDT | [1] Epoch -552 finished +---------------------------------- --------------- +epoch -552 +replay_buffer/size 999033 +trainer/num train calls 449000 +trainer/Policy Loss -20.0938 +trainer/Log Pis Mean 24.0437 +trainer/Log Pis Std 12.81 +trainer/Log Pis Max 80.338 +trainer/Log Pis Min -5.85571 +trainer/policy/mean Mean -0.0465099 +trainer/policy/mean Std 0.908858 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83582 +trainer/policy/normal/std Std 0.609822 +trainer/policy/normal/std Max 5.23645 +trainer/policy/normal/std Min 0.33333 +trainer/policy/normal/log_std Mean 1.01016 +trainer/policy/normal/log_std Std 0.2825 +trainer/policy/normal/log_std Max 1.65564 +trainer/policy/normal/log_std Min -1.09862 +eval/num steps total 447293 +eval/num paths total 449 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.182257 +eval/Actions Std 0.926235 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62711 +time/logging (s) 0.00371682 +time/sampling batch (s) 0.278804 +time/saving (s) 0.00342335 +time/training (s) 7.13618 +time/epoch (s) 10.0492 +time/total (s) 4673.83 +Epoch -552 +---------------------------------- --------------- +2022-05-10 14:28:45.201542 PDT | [1] Epoch -551 finished +---------------------------------- --------------- +epoch -551 +replay_buffer/size 999033 +trainer/num train calls 450000 +trainer/Policy Loss -20.7248 +trainer/Log Pis Mean 23.7032 +trainer/Log Pis Std 12.8627 +trainer/Log Pis Max 73.746 +trainer/Log Pis Min -9.96626 +trainer/policy/mean Mean -0.0398694 +trainer/policy/mean Std 0.905696 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78052 +trainer/policy/normal/std Std 0.644319 +trainer/policy/normal/std Max 4.83812 +trainer/policy/normal/std Min 0.257555 +trainer/policy/normal/log_std Mean 0.981895 +trainer/policy/normal/log_std Std 0.327267 +trainer/policy/normal/log_std Max 1.57653 +trainer/policy/normal/log_std Min -1.35652 +eval/num steps total 448293 +eval/num paths total 450 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.230533 +eval/Actions Std 0.89107 +eval/Actions Max 0.999935 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4015 +time/logging (s) 0.00370894 +time/sampling batch (s) 0.282909 +time/saving (s) 0.00348788 +time/training (s) 7.17873 +time/epoch (s) 9.87033 +time/total (s) 4683.7 +Epoch -551 +---------------------------------- --------------- +2022-05-10 14:28:55.872109 PDT | [1] Epoch -550 finished +---------------------------------- --------------- +epoch -550 +replay_buffer/size 999033 +trainer/num train calls 451000 +trainer/Policy Loss -19.4485 +trainer/Log Pis Mean 23.391 +trainer/Log Pis Std 13.0947 +trainer/Log Pis Max 65.3928 +trainer/Log Pis Min -4.74606 +trainer/policy/mean Mean 0.00316434 +trainer/policy/mean Std 0.903947 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78546 +trainer/policy/normal/std Std 0.644115 +trainer/policy/normal/std Max 5.37914 +trainer/policy/normal/std Min 0.363322 +trainer/policy/normal/log_std Mean 0.986363 +trainer/policy/normal/log_std Std 0.30968 +trainer/policy/normal/log_std Max 1.68253 +trainer/policy/normal/log_std Min -1.01247 +eval/num steps total 449293 +eval/num paths total 451 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.159358 +eval/Actions Std 0.942944 +eval/Actions Max 0.999998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55837 +time/logging (s) 0.00373185 +time/sampling batch (s) 0.52723 +time/saving (s) 0.00351218 +time/training (s) 7.55554 +time/epoch (s) 10.6484 +time/total (s) 4694.36 +Epoch -550 +---------------------------------- --------------- +2022-05-10 14:29:05.503695 PDT | [1] Epoch -549 finished +---------------------------------- --------------- +epoch -549 +replay_buffer/size 999033 +trainer/num train calls 452000 +trainer/Policy Loss -20.0612 +trainer/Log Pis Mean 23.9591 +trainer/Log Pis Std 12.8047 +trainer/Log Pis Max 64.9823 +trainer/Log Pis Min -6.8949 +trainer/policy/mean Mean -0.0243604 +trainer/policy/mean Std 0.908012 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80446 +trainer/policy/normal/std Std 0.653552 +trainer/policy/normal/std Max 6.34708 +trainer/policy/normal/std Min 0.249739 +trainer/policy/normal/log_std Mean 0.992284 +trainer/policy/normal/log_std Std 0.31541 +trainer/policy/normal/log_std Max 1.84799 +trainer/policy/normal/log_std Min -1.38734 +eval/num steps total 450293 +eval/num paths total 452 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0730115 +eval/Actions Std 0.900661 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52187 +time/logging (s) 0.00411018 +time/sampling batch (s) 0.275518 +time/saving (s) 0.00395824 +time/training (s) 6.80438 +time/epoch (s) 9.60985 +time/total (s) 4703.97 +Epoch -549 +---------------------------------- --------------- +2022-05-10 14:29:16.610662 PDT | [1] Epoch -548 finished +---------------------------------- --------------- +epoch -548 +replay_buffer/size 999033 +trainer/num train calls 453000 +trainer/Policy Loss -19.8056 +trainer/Log Pis Mean 24.7034 +trainer/Log Pis Std 13.2567 +trainer/Log Pis Max 80.2001 +trainer/Log Pis Min -6.79366 +trainer/policy/mean Mean -0.0396219 +trainer/policy/mean Std 0.905455 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84147 +trainer/policy/normal/std Std 0.64964 +trainer/policy/normal/std Max 6.85548 +trainer/policy/normal/std Min 0.289482 +trainer/policy/normal/log_std Mean 1.00582 +trainer/policy/normal/log_std Std 0.315033 +trainer/policy/normal/log_std Max 1.92505 +trainer/policy/normal/log_std Min -1.23966 +eval/num steps total 451293 +eval/num paths total 453 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.146473 +eval/Actions Std 0.884672 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63551 +time/logging (s) 0.0037537 +time/sampling batch (s) 0.277791 +time/saving (s) 0.00356464 +time/training (s) 8.16352 +time/epoch (s) 11.0841 +time/total (s) 4715.06 +Epoch -548 +---------------------------------- --------------- +2022-05-10 14:29:28.178925 PDT | [1] Epoch -547 finished +---------------------------------- --------------- +epoch -547 +replay_buffer/size 999033 +trainer/num train calls 454000 +trainer/Policy Loss -19.2048 +trainer/Log Pis Mean 25.1784 +trainer/Log Pis Std 13.0654 +trainer/Log Pis Max 81.4808 +trainer/Log Pis Min -5.64024 +trainer/policy/mean Mean -0.0622462 +trainer/policy/mean Std 0.906154 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84047 +trainer/policy/normal/std Std 0.639371 +trainer/policy/normal/std Max 5.38585 +trainer/policy/normal/std Min 0.321322 +trainer/policy/normal/log_std Mean 1.00772 +trainer/policy/normal/log_std Std 0.302373 +trainer/policy/normal/log_std Max 1.68378 +trainer/policy/normal/log_std Min -1.13531 +eval/num steps total 452293 +eval/num paths total 454 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.50558 +eval/Actions Std 0.753571 +eval/Actions Max 0.999995 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72423 +time/logging (s) 0.00367407 +time/sampling batch (s) 0.277165 +time/saving (s) 0.00343774 +time/training (s) 8.5374 +time/epoch (s) 11.5459 +time/total (s) 4726.61 +Epoch -547 +---------------------------------- --------------- +2022-05-10 14:29:37.479290 PDT | [1] Epoch -546 finished +---------------------------------- --------------- +epoch -546 +replay_buffer/size 999033 +trainer/num train calls 455000 +trainer/Policy Loss -20.5203 +trainer/Log Pis Mean 24.1921 +trainer/Log Pis Std 13.2108 +trainer/Log Pis Max 69.406 +trainer/Log Pis Min -8.63045 +trainer/policy/mean Mean -0.0238876 +trainer/policy/mean Std 0.911218 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74826 +trainer/policy/normal/std Std 0.62626 +trainer/policy/normal/std Max 5.97209 +trainer/policy/normal/std Min 0.272528 +trainer/policy/normal/log_std Mean 0.973451 +trainer/policy/normal/log_std Std 0.309451 +trainer/policy/normal/log_std Max 1.7871 +trainer/policy/normal/log_std Min -1.30001 +eval/num steps total 453293 +eval/num paths total 455 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104932 +eval/Actions Std 0.916162 +eval/Actions Max 0.999985 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64907 +time/logging (s) 0.00368251 +time/sampling batch (s) 0.280901 +time/saving (s) 0.00352786 +time/training (s) 6.34084 +time/epoch (s) 9.27802 +time/total (s) 4735.89 +Epoch -546 +---------------------------------- --------------- +2022-05-10 14:29:47.432980 PDT | [1] Epoch -545 finished +---------------------------------- --------------- +epoch -545 +replay_buffer/size 999033 +trainer/num train calls 456000 +trainer/Policy Loss -18.4774 +trainer/Log Pis Mean 24.054 +trainer/Log Pis Std 13.0663 +trainer/Log Pis Max 64.9189 +trainer/Log Pis Min -6.9609 +trainer/policy/mean Mean -0.0291219 +trainer/policy/mean Std 0.903732 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79062 +trainer/policy/normal/std Std 0.646047 +trainer/policy/normal/std Max 6.98302 +trainer/policy/normal/std Min 0.261026 +trainer/policy/normal/log_std Mean 0.987263 +trainer/policy/normal/log_std Std 0.316173 +trainer/policy/normal/log_std Max 1.94348 +trainer/policy/normal/log_std Min -1.34314 +eval/num steps total 454293 +eval/num paths total 456 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.31958 +eval/Actions Std 0.830057 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61281 +time/logging (s) 0.00376664 +time/sampling batch (s) 0.529023 +time/saving (s) 0.00345522 +time/training (s) 6.78249 +time/epoch (s) 9.93154 +time/total (s) 4745.82 +Epoch -545 +---------------------------------- --------------- +2022-05-10 14:29:57.461367 PDT | [1] Epoch -544 finished +---------------------------------- --------------- +epoch -544 +replay_buffer/size 999033 +trainer/num train calls 457000 +trainer/Policy Loss -19.6683 +trainer/Log Pis Mean 23.7302 +trainer/Log Pis Std 13.1276 +trainer/Log Pis Max 67.176 +trainer/Log Pis Min -8.49355 +trainer/policy/mean Mean -0.035501 +trainer/policy/mean Std 0.899804 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81284 +trainer/policy/normal/std Std 0.669851 +trainer/policy/normal/std Max 5.38524 +trainer/policy/normal/std Min 0.258573 +trainer/policy/normal/log_std Mean 0.991256 +trainer/policy/normal/log_std Std 0.33573 +trainer/policy/normal/log_std Max 1.68366 +trainer/policy/normal/log_std Min -1.35258 +eval/num steps total 455293 +eval/num paths total 457 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.367439 +eval/Actions Std 0.843437 +eval/Actions Max 0.999992 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53655 +time/logging (s) 0.00479499 +time/sampling batch (s) 0.288686 +time/saving (s) 0.00451672 +time/training (s) 7.17205 +time/epoch (s) 10.0066 +time/total (s) 4755.83 +Epoch -544 +---------------------------------- --------------- +2022-05-10 14:30:07.612248 PDT | [1] Epoch -543 finished +---------------------------------- --------------- +epoch -543 +replay_buffer/size 999033 +trainer/num train calls 458000 +trainer/Policy Loss -19.306 +trainer/Log Pis Mean 24.284 +trainer/Log Pis Std 13.4286 +trainer/Log Pis Max 66.3186 +trainer/Log Pis Min -9.19321 +trainer/policy/mean Mean -0.0279335 +trainer/policy/mean Std 0.904968 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8562 +trainer/policy/normal/std Std 0.658994 +trainer/policy/normal/std Max 5.83532 +trainer/policy/normal/std Min 0.306986 +trainer/policy/normal/log_std Mean 1.01019 +trainer/policy/normal/log_std Std 0.318558 +trainer/policy/normal/log_std Max 1.76393 +trainer/policy/normal/log_std Min -1.18095 +eval/num steps total 456293 +eval/num paths total 458 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.046357 +eval/Actions Std 0.896816 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57701 +time/logging (s) 0.00418568 +time/sampling batch (s) 0.298394 +time/saving (s) 0.00415145 +time/training (s) 7.24234 +time/epoch (s) 10.1261 +time/total (s) 4765.96 +Epoch -543 +---------------------------------- --------------- +2022-05-10 14:30:18.295946 PDT | [1] Epoch -542 finished +---------------------------------- --------------- +epoch -542 +replay_buffer/size 999033 +trainer/num train calls 459000 +trainer/Policy Loss -19.8624 +trainer/Log Pis Mean 24.4169 +trainer/Log Pis Std 13.0337 +trainer/Log Pis Max 66.9835 +trainer/Log Pis Min -6.90386 +trainer/policy/mean Mean -0.0513745 +trainer/policy/mean Std 0.908913 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86212 +trainer/policy/normal/std Std 0.64873 +trainer/policy/normal/std Max 6.39596 +trainer/policy/normal/std Min 0.327307 +trainer/policy/normal/log_std Mean 1.01351 +trainer/policy/normal/log_std Std 0.313805 +trainer/policy/normal/log_std Max 1.85567 +trainer/policy/normal/log_std Min -1.11686 +eval/num steps total 457293 +eval/num paths total 459 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0875906 +eval/Actions Std 0.902812 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6819 +time/logging (s) 0.00379087 +time/sampling batch (s) 0.283936 +time/saving (s) 0.00343907 +time/training (s) 7.68718 +time/epoch (s) 10.6602 +time/total (s) 4776.62 +Epoch -542 +---------------------------------- --------------- +2022-05-10 14:30:29.360335 PDT | [1] Epoch -541 finished +---------------------------------- --------------- +epoch -541 +replay_buffer/size 999033 +trainer/num train calls 460000 +trainer/Policy Loss -18.9406 +trainer/Log Pis Mean 23.6053 +trainer/Log Pis Std 13.6063 +trainer/Log Pis Max 71.125 +trainer/Log Pis Min -7.05613 +trainer/policy/mean Mean -0.0460635 +trainer/policy/mean Std 0.902066 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76028 +trainer/policy/normal/std Std 0.632117 +trainer/policy/normal/std Max 6.64672 +trainer/policy/normal/std Min 0.282678 +trainer/policy/normal/log_std Mean 0.977883 +trainer/policy/normal/log_std Std 0.30844 +trainer/policy/normal/log_std Max 1.89412 +trainer/policy/normal/log_std Min -1.26345 +eval/num steps total 458293 +eval/num paths total 460 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0978438 +eval/Actions Std 0.901804 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.604 +time/logging (s) 0.00371906 +time/sampling batch (s) 0.531239 +time/saving (s) 0.00345056 +time/training (s) 7.8994 +time/epoch (s) 11.0418 +time/total (s) 4787.67 +Epoch -541 +---------------------------------- --------------- +2022-05-10 14:30:39.413226 PDT | [1] Epoch -540 finished +---------------------------------- --------------- +epoch -540 +replay_buffer/size 999033 +trainer/num train calls 461000 +trainer/Policy Loss -19.3814 +trainer/Log Pis Mean 24.2058 +trainer/Log Pis Std 13.1942 +trainer/Log Pis Max 66.7773 +trainer/Log Pis Min -11.4276 +trainer/policy/mean Mean -0.0445188 +trainer/policy/mean Std 0.907206 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81304 +trainer/policy/normal/std Std 0.634818 +trainer/policy/normal/std Max 5.48335 +trainer/policy/normal/std Min 0.305968 +trainer/policy/normal/log_std Mean 0.997309 +trainer/policy/normal/log_std Std 0.307243 +trainer/policy/normal/log_std Max 1.70172 +trainer/policy/normal/log_std Min -1.18428 +eval/num steps total 459293 +eval/num paths total 461 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.23804 +eval/Actions Std 0.933859 +eval/Actions Max 0.999992 +eval/Actions Min -0.999932 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64166 +time/logging (s) 0.00407914 +time/sampling batch (s) 0.283075 +time/saving (s) 0.00396817 +time/training (s) 7.09781 +time/epoch (s) 10.0306 +time/total (s) 4797.7 +Epoch -540 +---------------------------------- --------------- +2022-05-10 14:30:50.011484 PDT | [1] Epoch -539 finished +---------------------------------- --------------- +epoch -539 +replay_buffer/size 999033 +trainer/num train calls 462000 +trainer/Policy Loss -20.0714 +trainer/Log Pis Mean 24.4179 +trainer/Log Pis Std 13.0586 +trainer/Log Pis Max 63.5354 +trainer/Log Pis Min -6.20036 +trainer/policy/mean Mean -0.0576055 +trainer/policy/mean Std 0.904067 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88332 +trainer/policy/normal/std Std 0.64962 +trainer/policy/normal/std Max 5.80066 +trainer/policy/normal/std Min 0.31599 +trainer/policy/normal/log_std Mean 1.02193 +trainer/policy/normal/log_std Std 0.30922 +trainer/policy/normal/log_std Max 1.75797 +trainer/policy/normal/log_std Min -1.15205 +eval/num steps total 460293 +eval/num paths total 462 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0103515 +eval/Actions Std 0.948989 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63956 +time/logging (s) 0.00421391 +time/sampling batch (s) 0.297804 +time/saving (s) 0.00415887 +time/training (s) 7.62904 +time/epoch (s) 10.5748 +time/total (s) 4808.28 +Epoch -539 +---------------------------------- --------------- +2022-05-10 14:31:01.411527 PDT | [1] Epoch -538 finished +---------------------------------- --------------- +epoch -538 +replay_buffer/size 999033 +trainer/num train calls 463000 +trainer/Policy Loss -20.2661 +trainer/Log Pis Mean 25.11 +trainer/Log Pis Std 12.7431 +trainer/Log Pis Max 72.3329 +trainer/Log Pis Min -3.89325 +trainer/policy/mean Mean -0.031125 +trainer/policy/mean Std 0.906357 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78457 +trainer/policy/normal/std Std 0.627991 +trainer/policy/normal/std Max 5.90429 +trainer/policy/normal/std Min 0.230486 +trainer/policy/normal/log_std Mean 0.987522 +trainer/policy/normal/log_std Std 0.305498 +trainer/policy/normal/log_std Max 1.77568 +trainer/policy/normal/log_std Min -1.46756 +eval/num steps total 461293 +eval/num paths total 463 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.142858 +eval/Actions Std 0.852197 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66985 +time/logging (s) 0.00391887 +time/sampling batch (s) 0.816942 +time/saving (s) 0.00363912 +time/training (s) 7.88223 +time/epoch (s) 11.3766 +time/total (s) 4819.66 +Epoch -538 +---------------------------------- --------------- +2022-05-10 14:31:11.904731 PDT | [1] Epoch -537 finished +---------------------------------- --------------- +epoch -537 +replay_buffer/size 999033 +trainer/num train calls 464000 +trainer/Policy Loss -20.1487 +trainer/Log Pis Mean 24.216 +trainer/Log Pis Std 13.1553 +trainer/Log Pis Max 75.195 +trainer/Log Pis Min -7.02332 +trainer/policy/mean Mean -0.0751535 +trainer/policy/mean Std 0.906074 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80229 +trainer/policy/normal/std Std 0.630962 +trainer/policy/normal/std Max 5.72326 +trainer/policy/normal/std Min 0.269618 +trainer/policy/normal/log_std Mean 0.993756 +trainer/policy/normal/log_std Std 0.306093 +trainer/policy/normal/log_std Max 1.74454 +trainer/policy/normal/log_std Min -1.31075 +eval/num steps total 462293 +eval/num paths total 464 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110982 +eval/Actions Std 0.914206 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63632 +time/logging (s) 0.00413194 +time/sampling batch (s) 0.528127 +time/saving (s) 0.00417351 +time/training (s) 7.29824 +time/epoch (s) 10.471 +time/total (s) 4830.13 +Epoch -537 +---------------------------------- --------------- +2022-05-10 14:31:21.362229 PDT | [1] Epoch -536 finished +---------------------------------- --------------- +epoch -536 +replay_buffer/size 999033 +trainer/num train calls 465000 +trainer/Policy Loss -19.2882 +trainer/Log Pis Mean 23.9436 +trainer/Log Pis Std 12.9862 +trainer/Log Pis Max 72.8559 +trainer/Log Pis Min -8.07792 +trainer/policy/mean Mean -0.033214 +trainer/policy/mean Std 0.907758 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81096 +trainer/policy/normal/std Std 0.658349 +trainer/policy/normal/std Max 5.64063 +trainer/policy/normal/std Min 0.250925 +trainer/policy/normal/log_std Mean 0.993626 +trainer/policy/normal/log_std Std 0.319971 +trainer/policy/normal/log_std Max 1.73 +trainer/policy/normal/log_std Min -1.3826 +eval/num steps total 463293 +eval/num paths total 465 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0949725 +eval/Actions Std 0.896088 +eval/Actions Max 0.999988 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66508 +time/logging (s) 0.00406519 +time/sampling batch (s) 0.279272 +time/saving (s) 0.00395452 +time/training (s) 6.48231 +time/epoch (s) 9.43468 +time/total (s) 4839.57 +Epoch -536 +---------------------------------- --------------- +2022-05-10 14:31:30.984641 PDT | [1] Epoch -535 finished +---------------------------------- --------------- +epoch -535 +replay_buffer/size 999033 +trainer/num train calls 466000 +trainer/Policy Loss -18.6418 +trainer/Log Pis Mean 24.4323 +trainer/Log Pis Std 13.3554 +trainer/Log Pis Max 70.1986 +trainer/Log Pis Min -8.36773 +trainer/policy/mean Mean -0.0385356 +trainer/policy/mean Std 0.902282 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77364 +trainer/policy/normal/std Std 0.644146 +trainer/policy/normal/std Max 6.07105 +trainer/policy/normal/std Min 0.361032 +trainer/policy/normal/log_std Mean 0.981173 +trainer/policy/normal/log_std Std 0.316337 +trainer/policy/normal/log_std Max 1.80353 +trainer/policy/normal/log_std Min -1.01879 +eval/num steps total 464293 +eval/num paths total 466 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0909298 +eval/Actions Std 0.913717 +eval/Actions Max 0.999989 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63914 +time/logging (s) 0.0041294 +time/sampling batch (s) 0.275584 +time/saving (s) 0.00391022 +time/training (s) 6.67732 +time/epoch (s) 9.60008 +time/total (s) 4849.18 +Epoch -535 +---------------------------------- --------------- +2022-05-10 14:31:40.625466 PDT | [1] Epoch -534 finished +---------------------------------- --------------- +epoch -534 +replay_buffer/size 999033 +trainer/num train calls 467000 +trainer/Policy Loss -18.7463 +trainer/Log Pis Mean 24.1577 +trainer/Log Pis Std 12.8568 +trainer/Log Pis Max 85.9524 +trainer/Log Pis Min -5.17223 +trainer/policy/mean Mean -0.016045 +trainer/policy/mean Std 0.906388 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80456 +trainer/policy/normal/std Std 0.640187 +trainer/policy/normal/std Max 5.65681 +trainer/policy/normal/std Min 0.294908 +trainer/policy/normal/log_std Mean 0.994898 +trainer/policy/normal/log_std Std 0.300626 +trainer/policy/normal/log_std Max 1.73286 +trainer/policy/normal/log_std Min -1.22109 +eval/num steps total 465293 +eval/num paths total 467 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.022254 +eval/Actions Std 0.908861 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63857 +time/logging (s) 0.00396772 +time/sampling batch (s) 0.276403 +time/saving (s) 0.00374301 +time/training (s) 6.69559 +time/epoch (s) 9.61827 +time/total (s) 4858.8 +Epoch -534 +---------------------------------- --------------- +2022-05-10 14:31:50.617765 PDT | [1] Epoch -533 finished +---------------------------------- --------------- +epoch -533 +replay_buffer/size 999033 +trainer/num train calls 468000 +trainer/Policy Loss -19.3521 +trainer/Log Pis Mean 23.7321 +trainer/Log Pis Std 13.6681 +trainer/Log Pis Max 66.6686 +trainer/Log Pis Min -11.6812 +trainer/policy/mean Mean -0.0575594 +trainer/policy/mean Std 0.901965 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82976 +trainer/policy/normal/std Std 0.644559 +trainer/policy/normal/std Max 6.04224 +trainer/policy/normal/std Min 0.345538 +trainer/policy/normal/log_std Mean 1.00341 +trainer/policy/normal/log_std Std 0.304042 +trainer/policy/normal/log_std Max 1.79878 +trainer/policy/normal/log_std Min -1.06265 +eval/num steps total 466293 +eval/num paths total 468 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100109 +eval/Actions Std 0.915021 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6728 +time/logging (s) 0.00384616 +time/sampling batch (s) 0.276273 +time/saving (s) 0.00352227 +time/training (s) 7.01362 +time/epoch (s) 9.97006 +time/total (s) 4868.77 +Epoch -533 +---------------------------------- --------------- +2022-05-10 14:32:00.194810 PDT | [1] Epoch -532 finished +---------------------------------- --------------- +epoch -532 +replay_buffer/size 999033 +trainer/num train calls 469000 +trainer/Policy Loss -19.9541 +trainer/Log Pis Mean 24.9371 +trainer/Log Pis Std 13.1554 +trainer/Log Pis Max 72.5958 +trainer/Log Pis Min -5.73392 +trainer/policy/mean Mean -0.0209356 +trainer/policy/mean Std 0.913821 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82585 +trainer/policy/normal/std Std 0.611905 +trainer/policy/normal/std Max 6.00014 +trainer/policy/normal/std Min 0.32116 +trainer/policy/normal/log_std Mean 1.00578 +trainer/policy/normal/log_std Std 0.287827 +trainer/policy/normal/log_std Max 1.79178 +trainer/policy/normal/log_std Min -1.13582 +eval/num steps total 467293 +eval/num paths total 469 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.121937 +eval/Actions Std 0.823646 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43048 +time/logging (s) 0.00420503 +time/sampling batch (s) 0.775705 +time/saving (s) 0.00393152 +time/training (s) 6.34102 +time/epoch (s) 9.55534 +time/total (s) 4878.33 +Epoch -532 +---------------------------------- --------------- +2022-05-10 14:32:09.616238 PDT | [1] Epoch -531 finished +---------------------------------- --------------- +epoch -531 +replay_buffer/size 999033 +trainer/num train calls 470000 +trainer/Policy Loss -18.7762 +trainer/Log Pis Mean 24.5925 +trainer/Log Pis Std 13.6455 +trainer/Log Pis Max 72.027 +trainer/Log Pis Min -4.99056 +trainer/policy/mean Mean -0.0230837 +trainer/policy/mean Std 0.909184 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79711 +trainer/policy/normal/std Std 0.644067 +trainer/policy/normal/std Max 5.76664 +trainer/policy/normal/std Min 0.236552 +trainer/policy/normal/log_std Mean 0.98952 +trainer/policy/normal/log_std Std 0.318808 +trainer/policy/normal/log_std Max 1.75209 +trainer/policy/normal/log_std Min -1.44159 +eval/num steps total 468293 +eval/num paths total 470 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.215436 +eval/Actions Std 0.822961 +eval/Actions Max 0.999979 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64854 +time/logging (s) 0.00397739 +time/sampling batch (s) 0.528501 +time/saving (s) 0.00368785 +time/training (s) 6.21379 +time/epoch (s) 9.3985 +time/total (s) 4887.73 +Epoch -531 +---------------------------------- --------------- +2022-05-10 14:32:19.167919 PDT | [1] Epoch -530 finished +---------------------------------- --------------- +epoch -530 +replay_buffer/size 999033 +trainer/num train calls 471000 +trainer/Policy Loss -19.3144 +trainer/Log Pis Mean 25.1746 +trainer/Log Pis Std 13.4338 +trainer/Log Pis Max 68.7038 +trainer/Log Pis Min -5.99988 +trainer/policy/mean Mean -0.0624647 +trainer/policy/mean Std 0.903834 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.77204 +trainer/policy/normal/std Std 0.618026 +trainer/policy/normal/std Max 5.25747 +trainer/policy/normal/std Min 0.295134 +trainer/policy/normal/log_std Mean 0.983673 +trainer/policy/normal/log_std Std 0.302245 +trainer/policy/normal/log_std Max 1.65965 +trainer/policy/normal/log_std Min -1.22033 +eval/num steps total 469293 +eval/num paths total 471 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.248561 +eval/Actions Std 0.81944 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42307 +time/logging (s) 0.00374825 +time/sampling batch (s) 0.279459 +time/saving (s) 0.00346195 +time/training (s) 6.81941 +time/epoch (s) 9.52915 +time/total (s) 4897.26 +Epoch -530 +---------------------------------- --------------- +2022-05-10 14:32:29.789314 PDT | [1] Epoch -529 finished +---------------------------------- --------------- +epoch -529 +replay_buffer/size 999033 +trainer/num train calls 472000 +trainer/Policy Loss -19.0335 +trainer/Log Pis Mean 23.4673 +trainer/Log Pis Std 13.8668 +trainer/Log Pis Max 80.3685 +trainer/Log Pis Min -9.92979 +trainer/policy/mean Mean -0.0420892 +trainer/policy/mean Std 0.902029 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81296 +trainer/policy/normal/std Std 0.674279 +trainer/policy/normal/std Max 6.39231 +trainer/policy/normal/std Min 0.306592 +trainer/policy/normal/log_std Mean 0.990803 +trainer/policy/normal/log_std Std 0.336881 +trainer/policy/normal/log_std Max 1.8551 +trainer/policy/normal/log_std Min -1.18224 +eval/num steps total 470293 +eval/num paths total 472 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0379721 +eval/Actions Std 0.904008 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3826 +time/logging (s) 0.00378016 +time/sampling batch (s) 0.281347 +time/saving (s) 0.00402557 +time/training (s) 7.92723 +time/epoch (s) 10.599 +time/total (s) 4907.86 +Epoch -529 +---------------------------------- --------------- +2022-05-10 14:32:39.723216 PDT | [1] Epoch -528 finished +---------------------------------- --------------- +epoch -528 +replay_buffer/size 999033 +trainer/num train calls 473000 +trainer/Policy Loss -20.0252 +trainer/Log Pis Mean 24.0646 +trainer/Log Pis Std 12.536 +trainer/Log Pis Max 64.8978 +trainer/Log Pis Min -6.79291 +trainer/policy/mean Mean -0.017182 +trainer/policy/mean Std 0.904277 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79005 +trainer/policy/normal/std Std 0.650985 +trainer/policy/normal/std Max 5.64217 +trainer/policy/normal/std Min 0.251019 +trainer/policy/normal/log_std Mean 0.985593 +trainer/policy/normal/log_std Std 0.323671 +trainer/policy/normal/log_std Max 1.73027 +trainer/policy/normal/log_std Min -1.38223 +eval/num steps total 471293 +eval/num paths total 473 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.099032 +eval/Actions Std 0.921405 +eval/Actions Max 0.999987 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75318 +time/logging (s) 0.00365132 +time/sampling batch (s) 0.289571 +time/saving (s) 0.00340249 +time/training (s) 6.86045 +time/epoch (s) 9.91025 +time/total (s) 4917.78 +Epoch -528 +---------------------------------- --------------- +2022-05-10 14:32:49.288360 PDT | [1] Epoch -527 finished +---------------------------------- --------------- +epoch -527 +replay_buffer/size 999033 +trainer/num train calls 474000 +trainer/Policy Loss -19.8791 +trainer/Log Pis Mean 25.1113 +trainer/Log Pis Std 13.8818 +trainer/Log Pis Max 69.4026 +trainer/Log Pis Min -10.5346 +trainer/policy/mean Mean -0.0326089 +trainer/policy/mean Std 0.911298 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84258 +trainer/policy/normal/std Std 0.636357 +trainer/policy/normal/std Max 6.1977 +trainer/policy/normal/std Min 0.349506 +trainer/policy/normal/log_std Mean 1.01014 +trainer/policy/normal/log_std Std 0.293465 +trainer/policy/normal/log_std Max 1.82418 +trainer/policy/normal/log_std Min -1.05123 +eval/num steps total 472293 +eval/num paths total 474 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115718 +eval/Actions Std 0.920002 +eval/Actions Max 0.999996 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.29519 +time/logging (s) 0.00417909 +time/sampling batch (s) 0.526105 +time/saving (s) 0.00379969 +time/training (s) 6.71427 +time/epoch (s) 9.54354 +time/total (s) 4927.32 +Epoch -527 +---------------------------------- --------------- +2022-05-10 14:32:59.567969 PDT | [1] Epoch -526 finished +---------------------------------- --------------- +epoch -526 +replay_buffer/size 999033 +trainer/num train calls 475000 +trainer/Policy Loss -20.4514 +trainer/Log Pis Mean 24.1307 +trainer/Log Pis Std 12.369 +trainer/Log Pis Max 58.8379 +trainer/Log Pis Min -5.58566 +trainer/policy/mean Mean -0.0364555 +trainer/policy/mean Std 0.905976 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91226 +trainer/policy/normal/std Std 0.648258 +trainer/policy/normal/std Max 5.66051 +trainer/policy/normal/std Min 0.345672 +trainer/policy/normal/log_std Mean 1.03457 +trainer/policy/normal/log_std Std 0.293139 +trainer/policy/normal/log_std Max 1.73351 +trainer/policy/normal/log_std Min -1.06226 +eval/num steps total 473293 +eval/num paths total 475 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0667479 +eval/Actions Std 0.911175 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62251 +time/logging (s) 0.00384073 +time/sampling batch (s) 0.276706 +time/saving (s) 0.00353132 +time/training (s) 7.35026 +time/epoch (s) 10.2569 +time/total (s) 4937.58 +Epoch -526 +---------------------------------- --------------- +2022-05-10 14:33:10.600071 PDT | [1] Epoch -525 finished +---------------------------------- --------------- +epoch -525 +replay_buffer/size 999033 +trainer/num train calls 476000 +trainer/Policy Loss -20.3254 +trainer/Log Pis Mean 24.7582 +trainer/Log Pis Std 12.7051 +trainer/Log Pis Max 71.0853 +trainer/Log Pis Min -6.35645 +trainer/policy/mean Mean -0.0288934 +trainer/policy/mean Std 0.912212 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.79315 +trainer/policy/normal/std Std 0.635895 +trainer/policy/normal/std Max 6.74401 +trainer/policy/normal/std Min 0.34994 +trainer/policy/normal/log_std Mean 0.990338 +trainer/policy/normal/log_std Std 0.305061 +trainer/policy/normal/log_std Max 1.90865 +trainer/policy/normal/log_std Min -1.04999 +eval/num steps total 474293 +eval/num paths total 476 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118735 +eval/Actions Std 0.909014 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81236 +time/logging (s) 0.00397372 +time/sampling batch (s) 0.277342 +time/saving (s) 0.00358384 +time/training (s) 7.91275 +time/epoch (s) 11.01 +time/total (s) 4948.6 +Epoch -525 +---------------------------------- --------------- +2022-05-10 14:33:20.393271 PDT | [1] Epoch -524 finished +---------------------------------- --------------- +epoch -524 +replay_buffer/size 999033 +trainer/num train calls 477000 +trainer/Policy Loss -19.1786 +trainer/Log Pis Mean 24.7343 +trainer/Log Pis Std 12.7794 +trainer/Log Pis Max 69.8656 +trainer/Log Pis Min -12.5214 +trainer/policy/mean Mean -0.0301349 +trainer/policy/mean Std 0.90511 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80193 +trainer/policy/normal/std Std 0.642268 +trainer/policy/normal/std Max 5.21209 +trainer/policy/normal/std Min 0.318151 +trainer/policy/normal/log_std Mean 0.992894 +trainer/policy/normal/log_std Std 0.30678 +trainer/policy/normal/log_std Max 1.65098 +trainer/policy/normal/log_std Min -1.14523 +eval/num steps total 475293 +eval/num paths total 477 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.450348 +eval/Actions Std 0.847685 +eval/Actions Max 0.999978 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59941 +time/logging (s) 0.00377483 +time/sampling batch (s) 0.528399 +time/saving (s) 0.00345711 +time/training (s) 6.63549 +time/epoch (s) 9.77054 +time/total (s) 4958.37 +Epoch -524 +---------------------------------- --------------- +2022-05-10 14:33:30.849105 PDT | [1] Epoch -523 finished +---------------------------------- --------------- +epoch -523 +replay_buffer/size 999033 +trainer/num train calls 478000 +trainer/Policy Loss -18.6928 +trainer/Log Pis Mean 23.9125 +trainer/Log Pis Std 12.8795 +trainer/Log Pis Max 69.1035 +trainer/Log Pis Min -8.65764 +trainer/policy/mean Mean -0.0218224 +trainer/policy/mean Std 0.904493 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83271 +trainer/policy/normal/std Std 0.66303 +trainer/policy/normal/std Max 5.5233 +trainer/policy/normal/std Min 0.314894 +trainer/policy/normal/log_std Mean 1.00207 +trainer/policy/normal/log_std Std 0.315151 +trainer/policy/normal/log_std Max 1.70898 +trainer/policy/normal/log_std Min -1.15552 +eval/num steps total 476293 +eval/num paths total 478 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.204073 +eval/Actions Std 0.955678 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73028 +time/logging (s) 0.00373856 +time/sampling batch (s) 0.276302 +time/saving (s) 0.00346857 +time/training (s) 7.41992 +time/epoch (s) 10.4337 +time/total (s) 4968.81 +Epoch -523 +---------------------------------- --------------- +2022-05-10 14:33:41.122014 PDT | [1] Epoch -522 finished +---------------------------------- --------------- +epoch -522 +replay_buffer/size 999033 +trainer/num train calls 479000 +trainer/Policy Loss -18.7556 +trainer/Log Pis Mean 24.0279 +trainer/Log Pis Std 13.4355 +trainer/Log Pis Max 65.8174 +trainer/Log Pis Min -8.75704 +trainer/policy/mean Mean -0.0313408 +trainer/policy/mean Std 0.903886 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80593 +trainer/policy/normal/std Std 0.634509 +trainer/policy/normal/std Max 5.86139 +trainer/policy/normal/std Min 0.321515 +trainer/policy/normal/log_std Mean 0.99608 +trainer/policy/normal/log_std Std 0.298022 +trainer/policy/normal/log_std Max 1.76839 +trainer/policy/normal/log_std Min -1.13471 +eval/num steps total 477293 +eval/num paths total 479 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0849162 +eval/Actions Std 0.909702 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57906 +time/logging (s) 0.007064 +time/sampling batch (s) 0.277691 +time/saving (s) 0.00564091 +time/training (s) 7.38437 +time/epoch (s) 10.2538 +time/total (s) 4979.06 +Epoch -522 +---------------------------------- --------------- +2022-05-10 14:33:52.185310 PDT | [1] Epoch -521 finished +---------------------------------- --------------- +epoch -521 +replay_buffer/size 999033 +trainer/num train calls 480000 +trainer/Policy Loss -20.9773 +trainer/Log Pis Mean 25.1416 +trainer/Log Pis Std 12.5104 +trainer/Log Pis Max 66.7896 +trainer/Log Pis Min -10.8323 +trainer/policy/mean Mean -0.029954 +trainer/policy/mean Std 0.908951 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.857 +trainer/policy/normal/std Std 0.64271 +trainer/policy/normal/std Max 5.96609 +trainer/policy/normal/std Min 0.333757 +trainer/policy/normal/log_std Mean 1.01355 +trainer/policy/normal/log_std Std 0.303722 +trainer/policy/normal/log_std Max 1.78609 +trainer/policy/normal/log_std Min -1.09734 +eval/num steps total 478293 +eval/num paths total 480 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00762985 +eval/Actions Std 0.898712 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76106 +time/logging (s) 0.00396797 +time/sampling batch (s) 0.284441 +time/saving (s) 0.00374773 +time/training (s) 7.98059 +time/epoch (s) 11.0338 +time/total (s) 4990.1 +Epoch -521 +---------------------------------- --------------- +2022-05-10 14:34:02.435138 PDT | [1] Epoch -520 finished +---------------------------------- --------------- +epoch -520 +replay_buffer/size 999033 +trainer/num train calls 481000 +trainer/Policy Loss -19.2205 +trainer/Log Pis Mean 24.8232 +trainer/Log Pis Std 12.2555 +trainer/Log Pis Max 63.392 +trainer/Log Pis Min -3.70179 +trainer/policy/mean Mean -0.0386649 +trainer/policy/mean Std 0.903603 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82935 +trainer/policy/normal/std Std 0.671908 +trainer/policy/normal/std Max 6.17527 +trainer/policy/normal/std Min 0.271304 +trainer/policy/normal/log_std Mean 0.998726 +trainer/policy/normal/log_std Std 0.325429 +trainer/policy/normal/log_std Max 1.82055 +trainer/policy/normal/log_std Min -1.30452 +eval/num steps total 479293 +eval/num paths total 481 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.489942 +eval/Actions Std 0.827146 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65604 +time/logging (s) 0.00373691 +time/sampling batch (s) 0.783248 +time/saving (s) 0.00342855 +time/training (s) 6.78042 +time/epoch (s) 10.2269 +time/total (s) 5000.33 +Epoch -520 +---------------------------------- --------------- +2022-05-10 14:34:12.073497 PDT | [1] Epoch -519 finished +---------------------------------- --------------- +epoch -519 +replay_buffer/size 999033 +trainer/num train calls 482000 +trainer/Policy Loss -20.4834 +trainer/Log Pis Mean 25.487 +trainer/Log Pis Std 13.7742 +trainer/Log Pis Max 76.4051 +trainer/Log Pis Min -8.02876 +trainer/policy/mean Mean -0.0232275 +trainer/policy/mean Std 0.906759 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84944 +trainer/policy/normal/std Std 0.597 +trainer/policy/normal/std Max 5.93615 +trainer/policy/normal/std Min 0.376083 +trainer/policy/normal/log_std Mean 1.01767 +trainer/policy/normal/log_std Std 0.268096 +trainer/policy/normal/log_std Max 1.78106 +trainer/policy/normal/log_std Min -0.977946 +eval/num steps total 480293 +eval/num paths total 482 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.153679 +eval/Actions Std 0.924994 +eval/Actions Max 0.999986 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73681 +time/logging (s) 0.00374148 +time/sampling batch (s) 0.282404 +time/saving (s) 0.00351104 +time/training (s) 6.5894 +time/epoch (s) 9.61586 +time/total (s) 5009.95 +Epoch -519 +---------------------------------- --------------- +2022-05-10 14:34:22.344711 PDT | [1] Epoch -518 finished +---------------------------------- --------------- +epoch -518 +replay_buffer/size 999033 +trainer/num train calls 483000 +trainer/Policy Loss -19.6291 +trainer/Log Pis Mean 23.6122 +trainer/Log Pis Std 13.3943 +trainer/Log Pis Max 70.1204 +trainer/Log Pis Min -6.94223 +trainer/policy/mean Mean -0.0225871 +trainer/policy/mean Std 0.904606 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85167 +trainer/policy/normal/std Std 0.66064 +trainer/policy/normal/std Max 5.7266 +trainer/policy/normal/std Min 0.229578 +trainer/policy/normal/log_std Mean 1.00847 +trainer/policy/normal/log_std Std 0.320457 +trainer/policy/normal/log_std Max 1.74512 +trainer/policy/normal/log_std Min -1.47151 +eval/num steps total 481293 +eval/num paths total 483 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00406444 +eval/Actions Std 0.871459 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72855 +time/logging (s) 0.00376085 +time/sampling batch (s) 0.282519 +time/saving (s) 0.00346054 +time/training (s) 7.23022 +time/epoch (s) 10.2485 +time/total (s) 5020.2 +Epoch -518 +---------------------------------- --------------- +2022-05-10 14:34:33.354220 PDT | [1] Epoch -517 finished +---------------------------------- --------------- +epoch -517 +replay_buffer/size 999033 +trainer/num train calls 484000 +trainer/Policy Loss -18.607 +trainer/Log Pis Mean 25.1823 +trainer/Log Pis Std 14.3582 +trainer/Log Pis Max 83.1348 +trainer/Log Pis Min -11.327 +trainer/policy/mean Mean -0.0257096 +trainer/policy/mean Std 0.905433 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82823 +trainer/policy/normal/std Std 0.651821 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.303658 +trainer/policy/normal/log_std Mean 1.00162 +trainer/policy/normal/log_std Std 0.3117 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19185 +eval/num steps total 482293 +eval/num paths total 484 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216563 +eval/Actions Std 0.809143 +eval/Actions Max 0.999988 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82971 +time/logging (s) 0.00410545 +time/sampling batch (s) 0.532039 +time/saving (s) 0.004033 +time/training (s) 7.61735 +time/epoch (s) 10.9872 +time/total (s) 5031.19 +Epoch -517 +---------------------------------- --------------- +2022-05-10 14:34:43.675155 PDT | [1] Epoch -516 finished +---------------------------------- --------------- +epoch -516 +replay_buffer/size 999033 +trainer/num train calls 485000 +trainer/Policy Loss -19.5916 +trainer/Log Pis Mean 23.1084 +trainer/Log Pis Std 12.5442 +trainer/Log Pis Max 67.8543 +trainer/Log Pis Min -4.74398 +trainer/policy/mean Mean -0.0157115 +trainer/policy/mean Std 0.906351 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82469 +trainer/policy/normal/std Std 0.636357 +trainer/policy/normal/std Max 6.90585 +trainer/policy/normal/std Min 0.309512 +trainer/policy/normal/log_std Mean 1.00282 +trainer/policy/normal/log_std Std 0.298278 +trainer/policy/normal/log_std Max 1.93237 +trainer/policy/normal/log_std Min -1.17276 +eval/num steps total 483293 +eval/num paths total 485 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00386387 +eval/Actions Std 0.909275 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51655 +time/logging (s) 0.00375664 +time/sampling batch (s) 0.530046 +time/saving (s) 0.00372487 +time/training (s) 7.24343 +time/epoch (s) 10.2975 +time/total (s) 5041.49 +Epoch -516 +---------------------------------- --------------- +2022-05-10 14:34:53.282372 PDT | [1] Epoch -515 finished +---------------------------------- --------------- +epoch -515 +replay_buffer/size 999033 +trainer/num train calls 486000 +trainer/Policy Loss -20.0268 +trainer/Log Pis Mean 23.7949 +trainer/Log Pis Std 13.1008 +trainer/Log Pis Max 69.6111 +trainer/Log Pis Min -5.93486 +trainer/policy/mean Mean -0.0399011 +trainer/policy/mean Std 0.904482 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81519 +trainer/policy/normal/std Std 0.643803 +trainer/policy/normal/std Max 5.83307 +trainer/policy/normal/std Min 0.269519 +trainer/policy/normal/log_std Mean 0.996762 +trainer/policy/normal/log_std Std 0.313579 +trainer/policy/normal/log_std Max 1.76354 +trainer/policy/normal/log_std Min -1.31112 +eval/num steps total 484293 +eval/num paths total 486 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.186863 +eval/Actions Std 0.812142 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62945 +time/logging (s) 0.00375955 +time/sampling batch (s) 0.273338 +time/saving (s) 0.00343312 +time/training (s) 6.67518 +time/epoch (s) 9.58516 +time/total (s) 5051.08 +Epoch -515 +---------------------------------- --------------- +2022-05-10 14:35:02.468681 PDT | [1] Epoch -514 finished +---------------------------------- --------------- +epoch -514 +replay_buffer/size 999033 +trainer/num train calls 487000 +trainer/Policy Loss -19.5642 +trainer/Log Pis Mean 24.2693 +trainer/Log Pis Std 13.653 +trainer/Log Pis Max 67.7305 +trainer/Log Pis Min -5.70191 +trainer/policy/mean Mean -0.049626 +trainer/policy/mean Std 0.911223 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79614 +trainer/policy/normal/std Std 0.607836 +trainer/policy/normal/std Max 5.20082 +trainer/policy/normal/std Min 0.297985 +trainer/policy/normal/log_std Mean 0.994675 +trainer/policy/normal/log_std Std 0.29108 +trainer/policy/normal/log_std Max 1.64882 +trainer/policy/normal/log_std Min -1.21071 +eval/num steps total 485293 +eval/num paths total 487 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.178706 +eval/Actions Std 0.788272 +eval/Actions Max 0.999986 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43461 +time/logging (s) 0.0037003 +time/sampling batch (s) 0.274062 +time/saving (s) 0.00344925 +time/training (s) 6.44838 +time/epoch (s) 9.1642 +time/total (s) 5060.25 +Epoch -514 +---------------------------------- --------------- +2022-05-10 14:35:12.318379 PDT | [1] Epoch -513 finished +---------------------------------- --------------- +epoch -513 +replay_buffer/size 999033 +trainer/num train calls 488000 +trainer/Policy Loss -19.1568 +trainer/Log Pis Mean 24.5085 +trainer/Log Pis Std 12.9115 +trainer/Log Pis Max 73.4869 +trainer/Log Pis Min -5.11277 +trainer/policy/mean Mean -0.0309102 +trainer/policy/mean Std 0.902135 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83211 +trainer/policy/normal/std Std 0.642836 +trainer/policy/normal/std Max 5.33736 +trainer/policy/normal/std Min 0.271135 +trainer/policy/normal/log_std Mean 1.00312 +trainer/policy/normal/log_std Std 0.312795 +trainer/policy/normal/log_std Max 1.67473 +trainer/policy/normal/log_std Min -1.30514 +eval/num steps total 486293 +eval/num paths total 488 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0272552 +eval/Actions Std 0.907838 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.25552 +time/logging (s) 0.00366461 +time/sampling batch (s) 0.525076 +time/saving (s) 0.0034489 +time/training (s) 7.03991 +time/epoch (s) 9.82762 +time/total (s) 5070.08 +Epoch -513 +---------------------------------- --------------- +2022-05-10 14:35:22.352031 PDT | [1] Epoch -512 finished +---------------------------------- --------------- +epoch -512 +replay_buffer/size 999033 +trainer/num train calls 489000 +trainer/Policy Loss -19.4224 +trainer/Log Pis Mean 23.7412 +trainer/Log Pis Std 13.555 +trainer/Log Pis Max 66.2618 +trainer/Log Pis Min -5.54256 +trainer/policy/mean Mean -0.0322005 +trainer/policy/mean Std 0.906715 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.81439 +trainer/policy/normal/std Std 0.651179 +trainer/policy/normal/std Max 5.4265 +trainer/policy/normal/std Min 0.322197 +trainer/policy/normal/log_std Mean 0.996183 +trainer/policy/normal/log_std Std 0.312888 +trainer/policy/normal/log_std Max 1.69129 +trainer/policy/normal/log_std Min -1.13259 +eval/num steps total 487293 +eval/num paths total 489 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123727 +eval/Actions Std 0.911338 +eval/Actions Max 0.999963 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49307 +time/logging (s) 0.00373375 +time/sampling batch (s) 0.27508 +time/saving (s) 0.00344683 +time/training (s) 7.23638 +time/epoch (s) 10.0117 +time/total (s) 5080.09 +Epoch -512 +---------------------------------- --------------- +2022-05-10 14:35:32.036793 PDT | [1] Epoch -511 finished +---------------------------------- --------------- +epoch -511 +replay_buffer/size 999033 +trainer/num train calls 490000 +trainer/Policy Loss -20.0326 +trainer/Log Pis Mean 25.3398 +trainer/Log Pis Std 13.5333 +trainer/Log Pis Max 75.7908 +trainer/Log Pis Min -8.70719 +trainer/policy/mean Mean -0.0491356 +trainer/policy/mean Std 0.908412 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81945 +trainer/policy/normal/std Std 0.626938 +trainer/policy/normal/std Max 5.74584 +trainer/policy/normal/std Min 0.277281 +trainer/policy/normal/log_std Mean 1.00188 +trainer/policy/normal/log_std Std 0.295208 +trainer/policy/normal/log_std Max 1.74848 +trainer/policy/normal/log_std Min -1.28272 +eval/num steps total 488293 +eval/num paths total 490 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0147886 +eval/Actions Std 0.906611 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47105 +time/logging (s) 0.00407966 +time/sampling batch (s) 0.275322 +time/saving (s) 0.00394163 +time/training (s) 6.90854 +time/epoch (s) 9.66294 +time/total (s) 5089.76 +Epoch -511 +---------------------------------- --------------- +2022-05-10 14:35:43.067998 PDT | [1] Epoch -510 finished +---------------------------------- --------------- +epoch -510 +replay_buffer/size 999033 +trainer/num train calls 491000 +trainer/Policy Loss -18.87 +trainer/Log Pis Mean 24.3864 +trainer/Log Pis Std 12.7832 +trainer/Log Pis Max 70.7173 +trainer/Log Pis Min -8.91163 +trainer/policy/mean Mean -0.0264573 +trainer/policy/mean Std 0.900251 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80455 +trainer/policy/normal/std Std 0.640502 +trainer/policy/normal/std Max 6.3699 +trainer/policy/normal/std Min 0.330388 +trainer/policy/normal/log_std Mean 0.993901 +trainer/policy/normal/log_std Std 0.307246 +trainer/policy/normal/log_std Max 1.85158 +trainer/policy/normal/log_std Min -1.10749 +eval/num steps total 489293 +eval/num paths total 491 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.173941 +eval/Actions Std 0.876939 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64175 +time/logging (s) 0.00368063 +time/sampling batch (s) 0.278121 +time/saving (s) 0.00353071 +time/training (s) 8.08101 +time/epoch (s) 11.0081 +time/total (s) 5100.77 +Epoch -510 +---------------------------------- --------------- +2022-05-10 14:35:53.383930 PDT | [1] Epoch -509 finished +---------------------------------- --------------- +epoch -509 +replay_buffer/size 999033 +trainer/num train calls 492000 +trainer/Policy Loss -19.0884 +trainer/Log Pis Mean 24.3785 +trainer/Log Pis Std 13.0552 +trainer/Log Pis Max 65.7175 +trainer/Log Pis Min -13.3671 +trainer/policy/mean Mean -0.0368516 +trainer/policy/mean Std 0.905876 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.86173 +trainer/policy/normal/std Std 0.658885 +trainer/policy/normal/std Max 7.06549 +trainer/policy/normal/std Min 0.32494 +trainer/policy/normal/log_std Mean 1.01373 +trainer/policy/normal/log_std Std 0.31034 +trainer/policy/normal/log_std Max 1.95522 +trainer/policy/normal/log_std Min -1.12411 +eval/num steps total 490293 +eval/num paths total 492 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.266094 +eval/Actions Std 0.833645 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72008 +time/logging (s) 0.00369403 +time/sampling batch (s) 0.775082 +time/saving (s) 0.0034662 +time/training (s) 6.79156 +time/epoch (s) 10.2939 +time/total (s) 5111.07 +Epoch -509 +---------------------------------- --------------- +2022-05-10 14:36:03.062015 PDT | [1] Epoch -508 finished +---------------------------------- --------------- +epoch -508 +replay_buffer/size 999033 +trainer/num train calls 493000 +trainer/Policy Loss -19.9963 +trainer/Log Pis Mean 23.8501 +trainer/Log Pis Std 12.9883 +trainer/Log Pis Max 73.6988 +trainer/Log Pis Min -11.1924 +trainer/policy/mean Mean -0.0147215 +trainer/policy/mean Std 0.904625 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82927 +trainer/policy/normal/std Std 0.652983 +trainer/policy/normal/std Max 5.25879 +trainer/policy/normal/std Min 0.355401 +trainer/policy/normal/log_std Mean 1.00228 +trainer/policy/normal/log_std Std 0.307593 +trainer/policy/normal/log_std Max 1.6599 +trainer/policy/normal/log_std Min -1.03451 +eval/num steps total 491293 +eval/num paths total 493 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.157575 +eval/Actions Std 0.949602 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53281 +time/logging (s) 0.00377257 +time/sampling batch (s) 0.278889 +time/saving (s) 0.00342643 +time/training (s) 6.83683 +time/epoch (s) 9.65573 +time/total (s) 5120.72 +Epoch -508 +---------------------------------- --------------- +2022-05-10 14:36:13.651242 PDT | [1] Epoch -507 finished +---------------------------------- --------------- +epoch -507 +replay_buffer/size 999033 +trainer/num train calls 494000 +trainer/Policy Loss -19.6901 +trainer/Log Pis Mean 24.3801 +trainer/Log Pis Std 13.6342 +trainer/Log Pis Max 67.5279 +trainer/Log Pis Min -9.25544 +trainer/policy/mean Mean -0.0340094 +trainer/policy/mean Std 0.905957 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79693 +trainer/policy/normal/std Std 0.651959 +trainer/policy/normal/std Max 5.44073 +trainer/policy/normal/std Min 0.292855 +trainer/policy/normal/log_std Mean 0.989216 +trainer/policy/normal/log_std Std 0.316696 +trainer/policy/normal/log_std Max 1.69391 +trainer/policy/normal/log_std Min -1.22808 +eval/num steps total 492293 +eval/num paths total 494 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111169 +eval/Actions Std 0.910633 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74609 +time/logging (s) 0.003917 +time/sampling batch (s) 0.776451 +time/saving (s) 0.00358888 +time/training (s) 7.03718 +time/epoch (s) 10.5672 +time/total (s) 5131.29 +Epoch -507 +---------------------------------- --------------- +2022-05-10 14:36:23.717793 PDT | [1] Epoch -506 finished +---------------------------------- --------------- +epoch -506 +replay_buffer/size 999033 +trainer/num train calls 495000 +trainer/Policy Loss -19.829 +trainer/Log Pis Mean 24.3186 +trainer/Log Pis Std 13.5203 +trainer/Log Pis Max 69.2391 +trainer/Log Pis Min -9.75561 +trainer/policy/mean Mean -0.0626915 +trainer/policy/mean Std 0.909889 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7881 +trainer/policy/normal/std Std 0.609809 +trainer/policy/normal/std Max 5.86514 +trainer/policy/normal/std Min 0.336156 +trainer/policy/normal/log_std Mean 0.990858 +trainer/policy/normal/log_std Std 0.296098 +trainer/policy/normal/log_std Max 1.76903 +trainer/policy/normal/log_std Min -1.09018 +eval/num steps total 493293 +eval/num paths total 495 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11124 +eval/Actions Std 0.896921 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69871 +time/logging (s) 0.00417293 +time/sampling batch (s) 0.275455 +time/saving (s) 0.00384074 +time/training (s) 7.06199 +time/epoch (s) 10.0442 +time/total (s) 5141.34 +Epoch -506 +---------------------------------- --------------- +2022-05-10 14:36:33.723201 PDT | [1] Epoch -505 finished +---------------------------------- --------------- +epoch -505 +replay_buffer/size 999033 +trainer/num train calls 496000 +trainer/Policy Loss -18.8565 +trainer/Log Pis Mean 25.0774 +trainer/Log Pis Std 13.2866 +trainer/Log Pis Max 68.7593 +trainer/Log Pis Min -6.74366 +trainer/policy/mean Mean -0.066703 +trainer/policy/mean Std 0.907485 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79866 +trainer/policy/normal/std Std 0.651802 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.302748 +trainer/policy/normal/log_std Mean 0.989912 +trainer/policy/normal/log_std Std 0.316714 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19486 +eval/num steps total 494293 +eval/num paths total 496 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107497 +eval/Actions Std 0.885025 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54855 +time/logging (s) 0.00389253 +time/sampling batch (s) 0.27812 +time/saving (s) 0.00375783 +time/training (s) 7.1478 +time/epoch (s) 9.98212 +time/total (s) 5151.33 +Epoch -505 +---------------------------------- --------------- +2022-05-10 14:36:43.971292 PDT | [1] Epoch -504 finished +---------------------------------- --------------- +epoch -504 +replay_buffer/size 999033 +trainer/num train calls 497000 +trainer/Policy Loss -19.4579 +trainer/Log Pis Mean 23.3522 +trainer/Log Pis Std 13.3327 +trainer/Log Pis Max 67.1963 +trainer/Log Pis Min -7.80026 +trainer/policy/mean Mean -0.0620107 +trainer/policy/mean Std 0.902491 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77741 +trainer/policy/normal/std Std 0.637829 +trainer/policy/normal/std Max 5.10763 +trainer/policy/normal/std Min 0.332856 +trainer/policy/normal/log_std Mean 0.983328 +trainer/policy/normal/log_std Std 0.311614 +trainer/policy/normal/log_std Max 1.63074 +trainer/policy/normal/log_std Min -1.10004 +eval/num steps total 495293 +eval/num paths total 497 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0268821 +eval/Actions Std 0.912249 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50616 +time/logging (s) 0.00378649 +time/sampling batch (s) 0.27788 +time/saving (s) 0.00343893 +time/training (s) 7.4342 +time/epoch (s) 10.2255 +time/total (s) 5161.56 +Epoch -504 +---------------------------------- --------------- +2022-05-10 14:36:54.599405 PDT | [1] Epoch -503 finished +---------------------------------- --------------- +epoch -503 +replay_buffer/size 999033 +trainer/num train calls 498000 +trainer/Policy Loss -18.935 +trainer/Log Pis Mean 25.0515 +trainer/Log Pis Std 13.839 +trainer/Log Pis Max 71.6165 +trainer/Log Pis Min -6.42358 +trainer/policy/mean Mean -0.0460779 +trainer/policy/mean Std 0.908532 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83133 +trainer/policy/normal/std Std 0.619342 +trainer/policy/normal/std Max 5.8341 +trainer/policy/normal/std Min 0.328657 +trainer/policy/normal/log_std Mean 1.00633 +trainer/policy/normal/log_std Std 0.29594 +trainer/policy/normal/log_std Max 1.76372 +trainer/policy/normal/log_std Min -1.11274 +eval/num steps total 496293 +eval/num paths total 498 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0287622 +eval/Actions Std 0.905845 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60866 +time/logging (s) 0.00388873 +time/sampling batch (s) 0.282844 +time/saving (s) 0.00347441 +time/training (s) 7.70669 +time/epoch (s) 10.6056 +time/total (s) 5172.16 +Epoch -503 +---------------------------------- --------------- +2022-05-10 14:37:04.714131 PDT | [1] Epoch -502 finished +---------------------------------- --------------- +epoch -502 +replay_buffer/size 999033 +trainer/num train calls 499000 +trainer/Policy Loss -18.2804 +trainer/Log Pis Mean 24.4352 +trainer/Log Pis Std 13.7554 +trainer/Log Pis Max 72.1373 +trainer/Log Pis Min -5.43957 +trainer/policy/mean Mean -0.0373655 +trainer/policy/mean Std 0.90382 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82717 +trainer/policy/normal/std Std 0.657726 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.310299 +trainer/policy/normal/log_std Mean 0.998933 +trainer/policy/normal/log_std Std 0.323909 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.17022 +eval/num steps total 497293 +eval/num paths total 499 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.252751 +eval/Actions Std 0.91436 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54891 +time/logging (s) 0.00374535 +time/sampling batch (s) 0.282769 +time/saving (s) 0.00348027 +time/training (s) 7.25292 +time/epoch (s) 10.0918 +time/total (s) 5182.26 +Epoch -502 +---------------------------------- --------------- +2022-05-10 14:37:14.424254 PDT | [1] Epoch -501 finished +---------------------------------- --------------- +epoch -501 +replay_buffer/size 999033 +trainer/num train calls 500000 +trainer/Policy Loss -19.0131 +trainer/Log Pis Mean 24.7034 +trainer/Log Pis Std 13.6303 +trainer/Log Pis Max 71.4013 +trainer/Log Pis Min -6.88756 +trainer/policy/mean Mean -0.0430942 +trainer/policy/mean Std 0.902324 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8142 +trainer/policy/normal/std Std 0.661114 +trainer/policy/normal/std Max 6.09834 +trainer/policy/normal/std Min 0.242506 +trainer/policy/normal/log_std Mean 0.99405 +trainer/policy/normal/log_std Std 0.324252 +trainer/policy/normal/log_std Max 1.80802 +trainer/policy/normal/log_std Min -1.41673 +eval/num steps total 498293 +eval/num paths total 500 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.205622 +eval/Actions Std 0.930612 +eval/Actions Max 0.999991 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42256 +time/logging (s) 0.00369638 +time/sampling batch (s) 0.279781 +time/saving (s) 0.00640204 +time/training (s) 6.97514 +time/epoch (s) 9.68758 +time/total (s) 5191.95 +Epoch -501 +---------------------------------- --------------- +2022-05-10 14:37:24.244040 PDT | [1] Epoch -500 finished +---------------------------------- --------------- +epoch -500 +replay_buffer/size 999033 +trainer/num train calls 501000 +trainer/Policy Loss -19.4963 +trainer/Log Pis Mean 24.329 +trainer/Log Pis Std 12.8859 +trainer/Log Pis Max 66.8074 +trainer/Log Pis Min -7.49906 +trainer/policy/mean Mean -0.0305754 +trainer/policy/mean Std 0.90659 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.83594 +trainer/policy/normal/std Std 0.618032 +trainer/policy/normal/std Max 5.84183 +trainer/policy/normal/std Min 0.362348 +trainer/policy/normal/log_std Mean 1.01007 +trainer/policy/normal/log_std Std 0.281822 +trainer/policy/normal/log_std Max 1.76504 +trainer/policy/normal/log_std Min -1.01515 +eval/num steps total 499293 +eval/num paths total 501 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0910091 +eval/Actions Std 0.914319 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67809 +time/logging (s) 0.00400493 +time/sampling batch (s) 0.279157 +time/saving (s) 0.00400752 +time/training (s) 6.83223 +time/epoch (s) 9.79749 +time/total (s) 5201.75 +Epoch -500 +---------------------------------- --------------- +2022-05-10 14:37:33.979117 PDT | [1] Epoch -499 finished +---------------------------------- --------------- +epoch -499 +replay_buffer/size 999033 +trainer/num train calls 502000 +trainer/Policy Loss -19.8678 +trainer/Log Pis Mean 24.6976 +trainer/Log Pis Std 13.281 +trainer/Log Pis Max 74.2943 +trainer/Log Pis Min -7.29512 +trainer/policy/mean Mean -0.00942382 +trainer/policy/mean Std 0.903234 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.79669 +trainer/policy/normal/std Std 0.683979 +trainer/policy/normal/std Max 6.16556 +trainer/policy/normal/std Min 0.316264 +trainer/policy/normal/log_std Mean 0.984997 +trainer/policy/normal/log_std Std 0.333256 +trainer/policy/normal/log_std Max 1.81898 +trainer/policy/normal/log_std Min -1.15118 +eval/num steps total 500293 +eval/num paths total 502 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00527438 +eval/Actions Std 0.961742 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68493 +time/logging (s) 0.00379284 +time/sampling batch (s) 0.531262 +time/saving (s) 0.00352413 +time/training (s) 6.48839 +time/epoch (s) 9.7119 +time/total (s) 5211.47 +Epoch -499 +---------------------------------- --------------- +2022-05-10 14:37:43.514432 PDT | [1] Epoch -498 finished +---------------------------------- --------------- +epoch -498 +replay_buffer/size 999033 +trainer/num train calls 503000 +trainer/Policy Loss -19.0307 +trainer/Log Pis Mean 23.4006 +trainer/Log Pis Std 12.9813 +trainer/Log Pis Max 67.8344 +trainer/Log Pis Min -8.12186 +trainer/policy/mean Mean -0.0393989 +trainer/policy/mean Std 0.901751 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.7866 +trainer/policy/normal/std Std 0.637143 +trainer/policy/normal/std Max 5.54309 +trainer/policy/normal/std Min 0.30674 +trainer/policy/normal/log_std Mean 0.986938 +trainer/policy/normal/log_std Std 0.311121 +trainer/policy/normal/log_std Max 1.71255 +trainer/policy/normal/log_std Min -1.18176 +eval/num steps total 501293 +eval/num paths total 503 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232041 +eval/Actions Std 0.907458 +eval/Actions Max 0.999987 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76544 +time/logging (s) 0.00369254 +time/sampling batch (s) 0.53032 +time/saving (s) 0.00339928 +time/training (s) 6.20972 +time/epoch (s) 9.51257 +time/total (s) 5220.98 +Epoch -498 +---------------------------------- --------------- +2022-05-10 14:37:53.878760 PDT | [1] Epoch -497 finished +---------------------------------- --------------- +epoch -497 +replay_buffer/size 999033 +trainer/num train calls 504000 +trainer/Policy Loss -20.1168 +trainer/Log Pis Mean 24.6359 +trainer/Log Pis Std 13.5743 +trainer/Log Pis Max 72.3283 +trainer/Log Pis Min -8.67975 +trainer/policy/mean Mean -0.0352042 +trainer/policy/mean Std 0.90822 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81205 +trainer/policy/normal/std Std 0.646554 +trainer/policy/normal/std Max 6.26403 +trainer/policy/normal/std Min 0.339661 +trainer/policy/normal/log_std Mean 0.996219 +trainer/policy/normal/log_std Std 0.30787 +trainer/policy/normal/log_std Max 1.83482 +trainer/policy/normal/log_std Min -1.07981 +eval/num steps total 502293 +eval/num paths total 504 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.181663 +eval/Actions Std 0.904176 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36239 +time/logging (s) 0.00375098 +time/sampling batch (s) 0.530195 +time/saving (s) 0.00344712 +time/training (s) 7.4419 +time/epoch (s) 10.3417 +time/total (s) 5231.33 +Epoch -497 +---------------------------------- --------------- +2022-05-10 14:38:04.006113 PDT | [1] Epoch -496 finished +---------------------------------- --------------- +epoch -496 +replay_buffer/size 999033 +trainer/num train calls 505000 +trainer/Policy Loss -21.6655 +trainer/Log Pis Mean 24.2954 +trainer/Log Pis Std 13.5815 +trainer/Log Pis Max 64.431 +trainer/Log Pis Min -10.61 +trainer/policy/mean Mean -0.0373293 +trainer/policy/mean Std 0.911837 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80873 +trainer/policy/normal/std Std 0.657372 +trainer/policy/normal/std Max 7.04937 +trainer/policy/normal/std Min 0.24745 +trainer/policy/normal/log_std Mean 0.991722 +trainer/policy/normal/log_std Std 0.327244 +trainer/policy/normal/log_std Max 1.95294 +trainer/policy/normal/log_std Min -1.39654 +eval/num steps total 503293 +eval/num paths total 505 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.161112 +eval/Actions Std 0.862207 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40314 +time/logging (s) 0.00380446 +time/sampling batch (s) 0.532797 +time/saving (s) 0.00344724 +time/training (s) 7.16153 +time/epoch (s) 10.1047 +time/total (s) 5241.43 +Epoch -496 +---------------------------------- --------------- +2022-05-10 14:38:15.028251 PDT | [1] Epoch -495 finished +---------------------------------- --------------- +epoch -495 +replay_buffer/size 999033 +trainer/num train calls 506000 +trainer/Policy Loss -19.7641 +trainer/Log Pis Mean 23.8321 +trainer/Log Pis Std 12.7114 +trainer/Log Pis Max 64.0391 +trainer/Log Pis Min -4.70884 +trainer/policy/mean Mean -0.0602106 +trainer/policy/mean Std 0.903349 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86918 +trainer/policy/normal/std Std 0.642944 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.320104 +trainer/policy/normal/log_std Mean 1.01793 +trainer/policy/normal/log_std Std 0.303019 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.13911 +eval/num steps total 504293 +eval/num paths total 506 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0999708 +eval/Actions Std 0.914466 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61331 +time/logging (s) 0.00436966 +time/sampling batch (s) 0.283903 +time/saving (s) 0.00396269 +time/training (s) 8.09419 +time/epoch (s) 10.9997 +time/total (s) 5252.44 +Epoch -495 +---------------------------------- --------------- +2022-05-10 14:38:26.031810 PDT | [1] Epoch -494 finished +---------------------------------- --------------- +epoch -494 +replay_buffer/size 999033 +trainer/num train calls 507000 +trainer/Policy Loss -18.8656 +trainer/Log Pis Mean 24.0333 +trainer/Log Pis Std 12.2966 +trainer/Log Pis Max 66.8896 +trainer/Log Pis Min -7.90258 +trainer/policy/mean Mean -0.0217116 +trainer/policy/mean Std 0.905071 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.78056 +trainer/policy/normal/std Std 0.619938 +trainer/policy/normal/std Max 6.30778 +trainer/policy/normal/std Min 0.274909 +trainer/policy/normal/log_std Mean 0.986584 +trainer/policy/normal/log_std Std 0.303795 +trainer/policy/normal/log_std Max 1.84178 +trainer/policy/normal/log_std Min -1.29131 +eval/num steps total 505293 +eval/num paths total 507 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.566522 +eval/Actions Std 0.795223 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47857 +time/logging (s) 0.00376354 +time/sampling batch (s) 0.279433 +time/saving (s) 0.00362102 +time/training (s) 8.21456 +time/epoch (s) 10.98 +time/total (s) 5263.42 +Epoch -494 +---------------------------------- --------------- +2022-05-10 14:38:36.453147 PDT | [1] Epoch -493 finished +---------------------------------- --------------- +epoch -493 +replay_buffer/size 999033 +trainer/num train calls 508000 +trainer/Policy Loss -20.2106 +trainer/Log Pis Mean 23.936 +trainer/Log Pis Std 13.0345 +trainer/Log Pis Max 63.0858 +trainer/Log Pis Min -10.0916 +trainer/policy/mean Mean -0.0375764 +trainer/policy/mean Std 0.906504 +trainer/policy/mean Max 1 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7895 +trainer/policy/normal/std Std 0.641415 +trainer/policy/normal/std Max 7.2721 +trainer/policy/normal/std Min 0.301529 +trainer/policy/normal/log_std Mean 0.98856 +trainer/policy/normal/log_std Std 0.306496 +trainer/policy/normal/log_std Max 1.98404 +trainer/policy/normal/log_std Min -1.19889 +eval/num steps total 506293 +eval/num paths total 508 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.384208 +eval/Actions Std 0.820975 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48232 +time/logging (s) 0.00375542 +time/sampling batch (s) 0.285964 +time/saving (s) 0.00344701 +time/training (s) 7.62285 +time/epoch (s) 10.3983 +time/total (s) 5273.82 +Epoch -493 +---------------------------------- --------------- +2022-05-10 14:38:45.807206 PDT | [1] Epoch -492 finished +---------------------------------- --------------- +epoch -492 +replay_buffer/size 999033 +trainer/num train calls 509000 +trainer/Policy Loss -20.0684 +trainer/Log Pis Mean 25.4893 +trainer/Log Pis Std 12.7279 +trainer/Log Pis Max 70.2206 +trainer/Log Pis Min -11.4613 +trainer/policy/mean Mean -0.0441505 +trainer/policy/mean Std 0.910443 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82415 +trainer/policy/normal/std Std 0.659575 +trainer/policy/normal/std Max 5.64427 +trainer/policy/normal/std Min 0.364319 +trainer/policy/normal/log_std Mean 0.998339 +trainer/policy/normal/log_std Std 0.319731 +trainer/policy/normal/log_std Max 1.73064 +trainer/policy/normal/log_std Min -1.00972 +eval/num steps total 507293 +eval/num paths total 509 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0883366 +eval/Actions Std 0.879104 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54404 +time/logging (s) 0.00378374 +time/sampling batch (s) 0.279231 +time/saving (s) 0.00348053 +time/training (s) 6.50087 +time/epoch (s) 9.33141 +time/total (s) 5283.15 +Epoch -492 +---------------------------------- --------------- +2022-05-10 14:38:57.046708 PDT | [1] Epoch -491 finished +---------------------------------- --------------- +epoch -491 +replay_buffer/size 999033 +trainer/num train calls 510000 +trainer/Policy Loss -20.0334 +trainer/Log Pis Mean 24.4259 +trainer/Log Pis Std 13.5465 +trainer/Log Pis Max 80.9596 +trainer/Log Pis Min -10.8034 +trainer/policy/mean Mean -0.0112644 +trainer/policy/mean Std 0.904019 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88834 +trainer/policy/normal/std Std 0.692481 +trainer/policy/normal/std Max 5.91771 +trainer/policy/normal/std Min 0.329864 +trainer/policy/normal/log_std Mean 1.01797 +trainer/policy/normal/log_std Std 0.333422 +trainer/policy/normal/log_std Max 1.77795 +trainer/policy/normal/log_std Min -1.10908 +eval/num steps total 508293 +eval/num paths total 510 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.156671 +eval/Actions Std 0.938723 +eval/Actions Max 0.999981 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73639 +time/logging (s) 0.00370766 +time/sampling batch (s) 0.528157 +time/saving (s) 0.00339926 +time/training (s) 7.9453 +time/epoch (s) 11.217 +time/total (s) 5294.37 +Epoch -491 +---------------------------------- --------------- +2022-05-10 14:39:07.765331 PDT | [1] Epoch -490 finished +---------------------------------- --------------- +epoch -490 +replay_buffer/size 999033 +trainer/num train calls 511000 +trainer/Policy Loss -19.4286 +trainer/Log Pis Mean 24.7095 +trainer/Log Pis Std 14.2336 +trainer/Log Pis Max 86.5094 +trainer/Log Pis Min -9.90819 +trainer/policy/mean Mean -0.0715399 +trainer/policy/mean Std 0.906783 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.8215 +trainer/policy/normal/std Std 0.621991 +trainer/policy/normal/std Max 6.23053 +trainer/policy/normal/std Min 0.330269 +trainer/policy/normal/log_std Mean 1.00326 +trainer/policy/normal/log_std Std 0.291626 +trainer/policy/normal/log_std Max 1.82946 +trainer/policy/normal/log_std Min -1.10785 +eval/num steps total 509293 +eval/num paths total 511 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.247929 +eval/Actions Std 0.867326 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50988 +time/logging (s) 0.0040447 +time/sampling batch (s) 0.277955 +time/saving (s) 0.00371861 +time/training (s) 7.9008 +time/epoch (s) 10.6964 +time/total (s) 5305.07 +Epoch -490 +---------------------------------- --------------- +2022-05-10 14:39:19.232480 PDT | [1] Epoch -489 finished +---------------------------------- --------------- +epoch -489 +replay_buffer/size 999033 +trainer/num train calls 512000 +trainer/Policy Loss -18.6641 +trainer/Log Pis Mean 23.5502 +trainer/Log Pis Std 14.2009 +trainer/Log Pis Max 82.0549 +trainer/Log Pis Min -7.80376 +trainer/policy/mean Mean -0.0356463 +trainer/policy/mean Std 0.906388 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80398 +trainer/policy/normal/std Std 0.652813 +trainer/policy/normal/std Max 5.6071 +trainer/policy/normal/std Min 0.313182 +trainer/policy/normal/log_std Mean 0.991306 +trainer/policy/normal/log_std Std 0.319391 +trainer/policy/normal/log_std Max 1.72403 +trainer/policy/normal/log_std Min -1.16097 +eval/num steps total 510293 +eval/num paths total 512 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243186 +eval/Actions Std 0.900596 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33886 +time/logging (s) 0.00376524 +time/sampling batch (s) 0.275894 +time/saving (s) 0.00370443 +time/training (s) 8.82209 +time/epoch (s) 11.4443 +time/total (s) 5316.52 +Epoch -489 +---------------------------------- --------------- +2022-05-10 14:39:29.371068 PDT | [1] Epoch -488 finished +---------------------------------- --------------- +epoch -488 +replay_buffer/size 999033 +trainer/num train calls 513000 +trainer/Policy Loss -18.7638 +trainer/Log Pis Mean 24.538 +trainer/Log Pis Std 12.9036 +trainer/Log Pis Max 63.4705 +trainer/Log Pis Min -6.13792 +trainer/policy/mean Mean -0.0178919 +trainer/policy/mean Std 0.906059 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83031 +trainer/policy/normal/std Std 0.651591 +trainer/policy/normal/std Max 5.51465 +trainer/policy/normal/std Min 0.291508 +trainer/policy/normal/log_std Mean 1.00189 +trainer/policy/normal/log_std Std 0.31317 +trainer/policy/normal/log_std Max 1.70741 +trainer/policy/normal/log_std Min -1.23269 +eval/num steps total 511293 +eval/num paths total 513 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.206558 +eval/Actions Std 0.913135 +eval/Actions Max 0.999997 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60542 +time/logging (s) 0.00366538 +time/sampling batch (s) 0.274535 +time/saving (s) 0.0034102 +time/training (s) 7.22915 +time/epoch (s) 10.1162 +time/total (s) 5326.64 +Epoch -488 +---------------------------------- --------------- +2022-05-10 14:39:40.293741 PDT | [1] Epoch -487 finished +---------------------------------- ---------------- +epoch -487 +replay_buffer/size 999033 +trainer/num train calls 514000 +trainer/Policy Loss -19.622 +trainer/Log Pis Mean 24.897 +trainer/Log Pis Std 13.6037 +trainer/Log Pis Max 66.932 +trainer/Log Pis Min -11.1764 +trainer/policy/mean Mean -0.0255709 +trainer/policy/mean Std 0.910811 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.85648 +trainer/policy/normal/std Std 0.657518 +trainer/policy/normal/std Max 5.49342 +trainer/policy/normal/std Min 0.239206 +trainer/policy/normal/log_std Mean 1.01178 +trainer/policy/normal/log_std Std 0.310129 +trainer/policy/normal/log_std Max 1.70355 +trainer/policy/normal/log_std Min -1.43043 +eval/num steps total 512293 +eval/num paths total 514 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000847628 +eval/Actions Std 0.901632 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41999 +time/logging (s) 0.00431121 +time/sampling batch (s) 0.53019 +time/saving (s) 0.00341999 +time/training (s) 7.94279 +time/epoch (s) 10.9007 +time/total (s) 5337.54 +Epoch -487 +---------------------------------- ---------------- +2022-05-10 14:39:51.106250 PDT | [1] Epoch -486 finished +---------------------------------- --------------- +epoch -486 +replay_buffer/size 999033 +trainer/num train calls 515000 +trainer/Policy Loss -19.6962 +trainer/Log Pis Mean 24.0517 +trainer/Log Pis Std 13.3959 +trainer/Log Pis Max 68.1293 +trainer/Log Pis Min -10.2607 +trainer/policy/mean Mean -0.033383 +trainer/policy/mean Std 0.905906 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79525 +trainer/policy/normal/std Std 0.676803 +trainer/policy/normal/std Max 5.95588 +trainer/policy/normal/std Min 0.248168 +trainer/policy/normal/log_std Mean 0.984143 +trainer/policy/normal/log_std Std 0.337703 +trainer/policy/normal/log_std Max 1.78438 +trainer/policy/normal/log_std Min -1.39365 +eval/num steps total 513293 +eval/num paths total 515 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.368591 +eval/Actions Std 0.86657 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52127 +time/logging (s) 0.00383312 +time/sampling batch (s) 0.534211 +time/saving (s) 0.00333979 +time/training (s) 7.72613 +time/epoch (s) 10.7888 +time/total (s) 5348.34 +Epoch -486 +---------------------------------- --------------- +2022-05-10 14:40:01.385384 PDT | [1] Epoch -485 finished +---------------------------------- --------------- +epoch -485 +replay_buffer/size 999033 +trainer/num train calls 516000 +trainer/Policy Loss -20.3893 +trainer/Log Pis Mean 25.0114 +trainer/Log Pis Std 13.6431 +trainer/Log Pis Max 74.3871 +trainer/Log Pis Min -7.39263 +trainer/policy/mean Mean -0.0301617 +trainer/policy/mean Std 0.907164 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.74135 +trainer/policy/normal/std Std 0.66619 +trainer/policy/normal/std Max 5.72418 +trainer/policy/normal/std Min 0.260386 +trainer/policy/normal/log_std Mean 0.962776 +trainer/policy/normal/log_std Std 0.347451 +trainer/policy/normal/log_std Max 1.7447 +trainer/policy/normal/log_std Min -1.34559 +eval/num steps total 514293 +eval/num paths total 516 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0919825 +eval/Actions Std 0.898729 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56357 +time/logging (s) 0.00385507 +time/sampling batch (s) 0.276218 +time/saving (s) 0.00350377 +time/training (s) 7.40955 +time/epoch (s) 10.2567 +time/total (s) 5358.6 +Epoch -485 +---------------------------------- --------------- +2022-05-10 14:40:11.840864 PDT | [1] Epoch -484 finished +---------------------------------- --------------- +epoch -484 +replay_buffer/size 999033 +trainer/num train calls 517000 +trainer/Policy Loss -18.8202 +trainer/Log Pis Mean 24.5725 +trainer/Log Pis Std 12.7261 +trainer/Log Pis Max 72.5886 +trainer/Log Pis Min -8.42533 +trainer/policy/mean Mean -0.0414876 +trainer/policy/mean Std 0.907663 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78635 +trainer/policy/normal/std Std 0.613788 +trainer/policy/normal/std Max 6.0923 +trainer/policy/normal/std Min 0.313292 +trainer/policy/normal/log_std Mean 0.991555 +trainer/policy/normal/log_std Std 0.286412 +trainer/policy/normal/log_std Max 1.80703 +trainer/policy/normal/log_std Min -1.16062 +eval/num steps total 515293 +eval/num paths total 517 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135401 +eval/Actions Std 0.889161 +eval/Actions Max 0.999981 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65086 +time/logging (s) 0.00429267 +time/sampling batch (s) 0.527093 +time/saving (s) 0.00391156 +time/training (s) 7.24724 +time/epoch (s) 10.4334 +time/total (s) 5369.03 +Epoch -484 +---------------------------------- --------------- +2022-05-10 14:40:22.354025 PDT | [1] Epoch -483 finished +---------------------------------- --------------- +epoch -483 +replay_buffer/size 999033 +trainer/num train calls 518000 +trainer/Policy Loss -20.0786 +trainer/Log Pis Mean 24.9997 +trainer/Log Pis Std 14.0623 +trainer/Log Pis Max 79.9108 +trainer/Log Pis Min -8.75871 +trainer/policy/mean Mean -0.0129723 +trainer/policy/mean Std 0.912353 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8254 +trainer/policy/normal/std Std 0.631844 +trainer/policy/normal/std Max 6.00533 +trainer/policy/normal/std Min 0.237038 +trainer/policy/normal/log_std Mean 1.00342 +trainer/policy/normal/log_std Std 0.298492 +trainer/policy/normal/log_std Max 1.79265 +trainer/policy/normal/log_std Min -1.43953 +eval/num steps total 516293 +eval/num paths total 518 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0142334 +eval/Actions Std 0.959767 +eval/Actions Max 0.999998 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67395 +time/logging (s) 0.00383901 +time/sampling batch (s) 0.276682 +time/saving (s) 0.00351262 +time/training (s) 7.53184 +time/epoch (s) 10.4898 +time/total (s) 5379.53 +Epoch -483 +---------------------------------- --------------- +2022-05-10 14:40:32.206816 PDT | [1] Epoch -482 finished +---------------------------------- --------------- +epoch -482 +replay_buffer/size 999033 +trainer/num train calls 519000 +trainer/Policy Loss -20.6194 +trainer/Log Pis Mean 24.2186 +trainer/Log Pis Std 13.2093 +trainer/Log Pis Max 69.6565 +trainer/Log Pis Min -5.40427 +trainer/policy/mean Mean -0.0135653 +trainer/policy/mean Std 0.905352 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85563 +trainer/policy/normal/std Std 0.65726 +trainer/policy/normal/std Max 5.76299 +trainer/policy/normal/std Min 0.244351 +trainer/policy/normal/log_std Mean 1.00974 +trainer/policy/normal/log_std Std 0.322452 +trainer/policy/normal/log_std Max 1.75146 +trainer/policy/normal/log_std Min -1.40915 +eval/num steps total 517293 +eval/num paths total 519 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113451 +eval/Actions Std 0.919922 +eval/Actions Max 0.999967 +eval/Actions Min -0.999958 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62866 +time/logging (s) 0.00373561 +time/sampling batch (s) 0.776643 +time/saving (s) 0.00334261 +time/training (s) 6.41782 +time/epoch (s) 9.8302 +time/total (s) 5389.36 +Epoch -482 +---------------------------------- --------------- +2022-05-10 14:40:42.506181 PDT | [1] Epoch -481 finished +---------------------------------- --------------- +epoch -481 +replay_buffer/size 999033 +trainer/num train calls 520000 +trainer/Policy Loss -19.4256 +trainer/Log Pis Mean 24.3865 +trainer/Log Pis Std 12.6048 +trainer/Log Pis Max 62.317 +trainer/Log Pis Min -9.566 +trainer/policy/mean Mean -0.0451328 +trainer/policy/mean Std 0.907961 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80546 +trainer/policy/normal/std Std 0.614202 +trainer/policy/normal/std Max 5.20773 +trainer/policy/normal/std Min 0.30922 +trainer/policy/normal/log_std Mean 0.997474 +trainer/policy/normal/log_std Std 0.29311 +trainer/policy/normal/log_std Max 1.65014 +trainer/policy/normal/log_std Min -1.1737 +eval/num steps total 518293 +eval/num paths total 520 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0503744 +eval/Actions Std 0.922449 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48872 +time/logging (s) 0.00370148 +time/sampling batch (s) 0.526219 +time/saving (s) 0.00337362 +time/training (s) 7.25497 +time/epoch (s) 10.277 +time/total (s) 5399.64 +Epoch -481 +---------------------------------- --------------- +2022-05-10 14:40:52.769301 PDT | [1] Epoch -480 finished +---------------------------------- --------------- +epoch -480 +replay_buffer/size 999033 +trainer/num train calls 521000 +trainer/Policy Loss -20.1735 +trainer/Log Pis Mean 25.9079 +trainer/Log Pis Std 13.8717 +trainer/Log Pis Max 71.5888 +trainer/Log Pis Min -12.851 +trainer/policy/mean Mean -0.0336581 +trainer/policy/mean Std 0.911975 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.85325 +trainer/policy/normal/std Std 0.661645 +trainer/policy/normal/std Max 6.80323 +trainer/policy/normal/std Min 0.296673 +trainer/policy/normal/log_std Mean 1.00934 +trainer/policy/normal/log_std Std 0.31694 +trainer/policy/normal/log_std Max 1.9174 +trainer/policy/normal/log_std Min -1.21512 +eval/num steps total 519293 +eval/num paths total 521 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0774788 +eval/Actions Std 0.914393 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67469 +time/logging (s) 0.00364433 +time/sampling batch (s) 0.276063 +time/saving (s) 0.0034369 +time/training (s) 7.28284 +time/epoch (s) 10.2407 +time/total (s) 5409.88 +Epoch -480 +---------------------------------- --------------- +2022-05-10 14:41:03.058892 PDT | [1] Epoch -479 finished +---------------------------------- --------------- +epoch -479 +replay_buffer/size 999033 +trainer/num train calls 522000 +trainer/Policy Loss -18.727 +trainer/Log Pis Mean 24.7805 +trainer/Log Pis Std 13.8695 +trainer/Log Pis Max 96.2363 +trainer/Log Pis Min -7.07821 +trainer/policy/mean Mean -0.0488571 +trainer/policy/mean Std 0.904674 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86565 +trainer/policy/normal/std Std 0.673632 +trainer/policy/normal/std Max 7.10967 +trainer/policy/normal/std Min 0.301141 +trainer/policy/normal/log_std Mean 1.0131 +trainer/policy/normal/log_std Std 0.317523 +trainer/policy/normal/log_std Max 1.96146 +trainer/policy/normal/log_std Min -1.20018 +eval/num steps total 520293 +eval/num paths total 522 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0838718 +eval/Actions Std 0.919094 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43614 +time/logging (s) 0.00403139 +time/sampling batch (s) 0.778002 +time/saving (s) 0.00374657 +time/training (s) 7.0456 +time/epoch (s) 10.2675 +time/total (s) 5420.15 +Epoch -479 +---------------------------------- --------------- +2022-05-10 14:41:13.767718 PDT | [1] Epoch -478 finished +---------------------------------- --------------- +epoch -478 +replay_buffer/size 999033 +trainer/num train calls 523000 +trainer/Policy Loss -20.0573 +trainer/Log Pis Mean 24.817 +trainer/Log Pis Std 12.7872 +trainer/Log Pis Max 71.8361 +trainer/Log Pis Min -7.08868 +trainer/policy/mean Mean -0.0336174 +trainer/policy/mean Std 0.903725 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.82472 +trainer/policy/normal/std Std 0.663273 +trainer/policy/normal/std Max 5.7956 +trainer/policy/normal/std Min 0.276054 +trainer/policy/normal/log_std Mean 0.997653 +trainer/policy/normal/log_std Std 0.324409 +trainer/policy/normal/log_std Max 1.7571 +trainer/policy/normal/log_std Min -1.28716 +eval/num steps total 521293 +eval/num paths total 523 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.294235 +eval/Actions Std 0.852434 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70065 +time/logging (s) 0.00382912 +time/sampling batch (s) 0.529766 +time/saving (s) 0.00376419 +time/training (s) 7.44776 +time/epoch (s) 10.6858 +time/total (s) 5430.84 +Epoch -478 +---------------------------------- --------------- +2022-05-10 14:41:23.810442 PDT | [1] Epoch -477 finished +---------------------------------- --------------- +epoch -477 +replay_buffer/size 999033 +trainer/num train calls 524000 +trainer/Policy Loss -18.0848 +trainer/Log Pis Mean 23.5622 +trainer/Log Pis Std 12.795 +trainer/Log Pis Max 68.7771 +trainer/Log Pis Min -8.33987 +trainer/policy/mean Mean -0.0370161 +trainer/policy/mean Std 0.903449 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82105 +trainer/policy/normal/std Std 0.652777 +trainer/policy/normal/std Max 6.05718 +trainer/policy/normal/std Min 0.276287 +trainer/policy/normal/log_std Mean 0.999499 +trainer/policy/normal/log_std Std 0.307261 +trainer/policy/normal/log_std Max 1.80124 +trainer/policy/normal/log_std Min -1.28632 +eval/num steps total 522293 +eval/num paths total 524 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0863 +eval/Actions Std 0.9016 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60505 +time/logging (s) 0.00368666 +time/sampling batch (s) 0.528721 +time/saving (s) 0.00346021 +time/training (s) 6.87889 +time/epoch (s) 10.0198 +time/total (s) 5440.86 +Epoch -477 +---------------------------------- --------------- +2022-05-10 14:41:34.652605 PDT | [1] Epoch -476 finished +---------------------------------- --------------- +epoch -476 +replay_buffer/size 999033 +trainer/num train calls 525000 +trainer/Policy Loss -19.4485 +trainer/Log Pis Mean 25.1848 +trainer/Log Pis Std 13.5544 +trainer/Log Pis Max 69.5694 +trainer/Log Pis Min -9.97342 +trainer/policy/mean Mean -0.0578015 +trainer/policy/mean Std 0.90613 +trainer/policy/mean Max 0.999965 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79595 +trainer/policy/normal/std Std 0.664076 +trainer/policy/normal/std Max 7.08535 +trainer/policy/normal/std Min 0.243731 +trainer/policy/normal/log_std Mean 0.986148 +trainer/policy/normal/log_std Std 0.330731 +trainer/policy/normal/log_std Max 1.95803 +trainer/policy/normal/log_std Min -1.41169 +eval/num steps total 523293 +eval/num paths total 525 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0989013 +eval/Actions Std 0.902652 +eval/Actions Max 0.999978 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61045 +time/logging (s) 0.00374036 +time/sampling batch (s) 0.279032 +time/saving (s) 0.0034511 +time/training (s) 7.92295 +time/epoch (s) 10.8196 +time/total (s) 5451.69 +Epoch -476 +---------------------------------- --------------- +2022-05-10 14:41:45.598044 PDT | [1] Epoch -475 finished +---------------------------------- --------------- +epoch -475 +replay_buffer/size 999033 +trainer/num train calls 526000 +trainer/Policy Loss -19.7276 +trainer/Log Pis Mean 23.9757 +trainer/Log Pis Std 12.6801 +trainer/Log Pis Max 65.9154 +trainer/Log Pis Min -9.8626 +trainer/policy/mean Mean -0.0386491 +trainer/policy/mean Std 0.907167 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.74711 +trainer/policy/normal/std Std 0.655928 +trainer/policy/normal/std Max 5.986 +trainer/policy/normal/std Min 0.276512 +trainer/policy/normal/log_std Mean 0.969006 +trainer/policy/normal/log_std Std 0.326828 +trainer/policy/normal/log_std Max 1.78942 +trainer/policy/normal/log_std Min -1.2855 +eval/num steps total 524293 +eval/num paths total 526 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00340226 +eval/Actions Std 0.921648 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5268 +time/logging (s) 0.00395198 +time/sampling batch (s) 0.78245 +time/saving (s) 0.0035567 +time/training (s) 7.60622 +time/epoch (s) 10.923 +time/total (s) 5462.61 +Epoch -475 +---------------------------------- --------------- +2022-05-10 14:41:57.067255 PDT | [1] Epoch -474 finished +---------------------------------- --------------- +epoch -474 +replay_buffer/size 999033 +trainer/num train calls 527000 +trainer/Policy Loss -19.5462 +trainer/Log Pis Mean 23.1727 +trainer/Log Pis Std 13.2377 +trainer/Log Pis Max 71.3328 +trainer/Log Pis Min -8.61447 +trainer/policy/mean Mean -0.0314319 +trainer/policy/mean Std 0.908911 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81514 +trainer/policy/normal/std Std 0.654115 +trainer/policy/normal/std Max 6.42057 +trainer/policy/normal/std Min 0.302929 +trainer/policy/normal/log_std Mean 0.995618 +trainer/policy/normal/log_std Std 0.31852 +trainer/policy/normal/log_std Max 1.85951 +trainer/policy/normal/log_std Min -1.19426 +eval/num steps total 525293 +eval/num paths total 527 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100928 +eval/Actions Std 0.745144 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62982 +time/logging (s) 0.00403229 +time/sampling batch (s) 0.528999 +time/saving (s) 0.00384607 +time/training (s) 8.27976 +time/epoch (s) 11.4465 +time/total (s) 5474.06 +Epoch -474 +---------------------------------- --------------- +2022-05-10 14:42:07.788213 PDT | [1] Epoch -473 finished +---------------------------------- --------------- +epoch -473 +replay_buffer/size 999033 +trainer/num train calls 528000 +trainer/Policy Loss -19.6252 +trainer/Log Pis Mean 24.8611 +trainer/Log Pis Std 12.8933 +trainer/Log Pis Max 74.5783 +trainer/Log Pis Min -7.8079 +trainer/policy/mean Mean -0.0470539 +trainer/policy/mean Std 0.906378 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79945 +trainer/policy/normal/std Std 0.62826 +trainer/policy/normal/std Max 5.79417 +trainer/policy/normal/std Min 0.297937 +trainer/policy/normal/log_std Mean 0.993806 +trainer/policy/normal/log_std Std 0.300118 +trainer/policy/normal/log_std Max 1.75685 +trainer/policy/normal/log_std Min -1.21087 +eval/num steps total 526293 +eval/num paths total 528 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.302626 +eval/Actions Std 0.82372 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68333 +time/logging (s) 0.00390395 +time/sampling batch (s) 1.02759 +time/saving (s) 0.00371115 +time/training (s) 6.97927 +time/epoch (s) 10.6978 +time/total (s) 5484.76 +Epoch -473 +---------------------------------- --------------- +2022-05-10 14:42:17.813101 PDT | [1] Epoch -472 finished +---------------------------------- --------------- +epoch -472 +replay_buffer/size 999033 +trainer/num train calls 529000 +trainer/Policy Loss -19.6914 +trainer/Log Pis Mean 23.6764 +trainer/Log Pis Std 12.5694 +trainer/Log Pis Max 60.6942 +trainer/Log Pis Min -10.9774 +trainer/policy/mean Mean -0.0370081 +trainer/policy/mean Std 0.902595 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80233 +trainer/policy/normal/std Std 0.652968 +trainer/policy/normal/std Max 5.80163 +trainer/policy/normal/std Min 0.24295 +trainer/policy/normal/log_std Mean 0.991706 +trainer/policy/normal/log_std Std 0.314331 +trainer/policy/normal/log_std Max 1.75814 +trainer/policy/normal/log_std Min -1.4149 +eval/num steps total 527293 +eval/num paths total 529 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.443891 +eval/Actions Std 0.847804 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74163 +time/logging (s) 0.00378534 +time/sampling batch (s) 0.277521 +time/saving (s) 0.00344021 +time/training (s) 6.97574 +time/epoch (s) 10.0021 +time/total (s) 5494.77 +Epoch -472 +---------------------------------- --------------- +2022-05-10 14:42:27.843593 PDT | [1] Epoch -471 finished +---------------------------------- --------------- +epoch -471 +replay_buffer/size 999033 +trainer/num train calls 530000 +trainer/Policy Loss -19.2054 +trainer/Log Pis Mean 23.9429 +trainer/Log Pis Std 12.5278 +trainer/Log Pis Max 69.0249 +trainer/Log Pis Min -7.59783 +trainer/policy/mean Mean -0.0256496 +trainer/policy/mean Std 0.906247 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84278 +trainer/policy/normal/std Std 0.639073 +trainer/policy/normal/std Max 5.49152 +trainer/policy/normal/std Min 0.332896 +trainer/policy/normal/log_std Mean 1.00963 +trainer/policy/normal/log_std Std 0.295998 +trainer/policy/normal/log_std Max 1.7032 +trainer/policy/normal/log_std Min -1.09992 +eval/num steps total 528293 +eval/num paths total 530 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00165394 +eval/Actions Std 0.900682 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84818 +time/logging (s) 0.00366972 +time/sampling batch (s) 0.526491 +time/saving (s) 0.00341937 +time/training (s) 6.62623 +time/epoch (s) 10.008 +time/total (s) 5504.78 +Epoch -471 +---------------------------------- --------------- +2022-05-10 14:42:38.392321 PDT | [1] Epoch -470 finished +---------------------------------- --------------- +epoch -470 +replay_buffer/size 999033 +trainer/num train calls 531000 +trainer/Policy Loss -19.1376 +trainer/Log Pis Mean 24.2884 +trainer/Log Pis Std 13.0731 +trainer/Log Pis Max 66.405 +trainer/Log Pis Min -8.04327 +trainer/policy/mean Mean -0.0414179 +trainer/policy/mean Std 0.906408 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79141 +trainer/policy/normal/std Std 0.647027 +trainer/policy/normal/std Max 5.8658 +trainer/policy/normal/std Min 0.313269 +trainer/policy/normal/log_std Mean 0.988252 +trainer/policy/normal/log_std Std 0.311005 +trainer/policy/normal/log_std Max 1.76914 +trainer/policy/normal/log_std Min -1.16069 +eval/num steps total 529293 +eval/num paths total 531 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0420758 +eval/Actions Std 0.877133 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76235 +time/logging (s) 0.00377058 +time/sampling batch (s) 0.777323 +time/saving (s) 0.00347765 +time/training (s) 6.97958 +time/epoch (s) 10.5265 +time/total (s) 5515.31 +Epoch -470 +---------------------------------- --------------- +2022-05-10 14:42:49.101360 PDT | [1] Epoch -469 finished +---------------------------------- --------------- +epoch -469 +replay_buffer/size 999033 +trainer/num train calls 532000 +trainer/Policy Loss -20.6046 +trainer/Log Pis Mean 24.1925 +trainer/Log Pis Std 12.9944 +trainer/Log Pis Max 65.5774 +trainer/Log Pis Min -13.1666 +trainer/policy/mean Mean -0.0322475 +trainer/policy/mean Std 0.907298 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.76155 +trainer/policy/normal/std Std 0.650802 +trainer/policy/normal/std Max 5.94217 +trainer/policy/normal/std Min 0.261311 +trainer/policy/normal/log_std Mean 0.97402 +trainer/policy/normal/log_std Std 0.329327 +trainer/policy/normal/log_std Max 1.78207 +trainer/policy/normal/log_std Min -1.34205 +eval/num steps total 530293 +eval/num paths total 532 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0206771 +eval/Actions Std 0.96294 +eval/Actions Max 0.999983 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68173 +time/logging (s) 0.00371968 +time/sampling batch (s) 0.278037 +time/saving (s) 0.00346268 +time/training (s) 7.71967 +time/epoch (s) 10.6866 +time/total (s) 5526 +Epoch -469 +---------------------------------- --------------- +2022-05-10 14:43:00.068379 PDT | [1] Epoch -468 finished +---------------------------------- --------------- +epoch -468 +replay_buffer/size 999033 +trainer/num train calls 533000 +trainer/Policy Loss -19.7839 +trainer/Log Pis Mean 25.4575 +trainer/Log Pis Std 13.4249 +trainer/Log Pis Max 70.5317 +trainer/Log Pis Min -5.48997 +trainer/policy/mean Mean -0.0446826 +trainer/policy/mean Std 0.909692 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76584 +trainer/policy/normal/std Std 0.640083 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.333057 +trainer/policy/normal/log_std Mean 0.97786 +trainer/policy/normal/log_std Std 0.319041 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.09944 +eval/num steps total 531293 +eval/num paths total 533 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0204708 +eval/Actions Std 0.907905 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75142 +time/logging (s) 0.00413416 +time/sampling batch (s) 0.276601 +time/saving (s) 0.0038412 +time/training (s) 7.90872 +time/epoch (s) 10.9447 +time/total (s) 5536.95 +Epoch -468 +---------------------------------- --------------- +2022-05-10 14:43:12.018794 PDT | [1] Epoch -467 finished +---------------------------------- --------------- +epoch -467 +replay_buffer/size 999033 +trainer/num train calls 534000 +trainer/Policy Loss -18.6947 +trainer/Log Pis Mean 24.3154 +trainer/Log Pis Std 12.9073 +trainer/Log Pis Max 62.4222 +trainer/Log Pis Min -6.54768 +trainer/policy/mean Mean -0.035627 +trainer/policy/mean Std 0.907114 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76459 +trainer/policy/normal/std Std 0.655364 +trainer/policy/normal/std Max 6.54136 +trainer/policy/normal/std Min 0.358611 +trainer/policy/normal/log_std Mean 0.976102 +trainer/policy/normal/log_std Std 0.32293 +trainer/policy/normal/log_std Max 1.87814 +trainer/policy/normal/log_std Min -1.02552 +eval/num steps total 532293 +eval/num paths total 534 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0878565 +eval/Actions Std 0.917471 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37323 +time/logging (s) 0.00372161 +time/sampling batch (s) 0.27728 +time/saving (s) 0.00350943 +time/training (s) 9.0188 +time/epoch (s) 11.6765 +time/total (s) 5548.88 +Epoch -467 +---------------------------------- --------------- +2022-05-10 14:43:21.454875 PDT | [1] Epoch -466 finished +---------------------------------- --------------- +epoch -466 +replay_buffer/size 999033 +trainer/num train calls 535000 +trainer/Policy Loss -19.8316 +trainer/Log Pis Mean 23.8841 +trainer/Log Pis Std 13.5189 +trainer/Log Pis Max 68.7515 +trainer/Log Pis Min -8.1139 +trainer/policy/mean Mean -0.0383844 +trainer/policy/mean Std 0.904481 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.77879 +trainer/policy/normal/std Std 0.637339 +trainer/policy/normal/std Max 5.69513 +trainer/policy/normal/std Min 0.261675 +trainer/policy/normal/log_std Mean 0.98217 +trainer/policy/normal/log_std Std 0.322601 +trainer/policy/normal/log_std Max 1.73961 +trainer/policy/normal/log_std Min -1.34065 +eval/num steps total 533293 +eval/num paths total 535 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0535762 +eval/Actions Std 0.899467 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50119 +time/logging (s) 0.00374419 +time/sampling batch (s) 0.276536 +time/saving (s) 0.00348919 +time/training (s) 6.62866 +time/epoch (s) 9.41362 +time/total (s) 5558.29 +Epoch -466 +---------------------------------- --------------- +2022-05-10 14:43:32.513657 PDT | [1] Epoch -465 finished +---------------------------------- --------------- +epoch -465 +replay_buffer/size 999033 +trainer/num train calls 536000 +trainer/Policy Loss -19.3942 +trainer/Log Pis Mean 24.7567 +trainer/Log Pis Std 12.5135 +trainer/Log Pis Max 64.7514 +trainer/Log Pis Min -8.20014 +trainer/policy/mean Mean -0.0428989 +trainer/policy/mean Std 0.907792 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85533 +trainer/policy/normal/std Std 0.659454 +trainer/policy/normal/std Max 6.82432 +trainer/policy/normal/std Min 0.25025 +trainer/policy/normal/log_std Mean 1.01143 +trainer/policy/normal/log_std Std 0.310152 +trainer/policy/normal/log_std Max 1.92049 +trainer/policy/normal/log_std Min -1.38529 +eval/num steps total 534293 +eval/num paths total 536 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00679182 +eval/Actions Std 0.857485 +eval/Actions Max 0.999993 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76508 +time/logging (s) 0.00365578 +time/sampling batch (s) 0.277488 +time/saving (s) 0.00343422 +time/training (s) 7.98631 +time/epoch (s) 11.036 +time/total (s) 5569.33 +Epoch -465 +---------------------------------- --------------- +2022-05-10 14:43:42.750020 PDT | [1] Epoch -464 finished +---------------------------------- --------------- +epoch -464 +replay_buffer/size 999033 +trainer/num train calls 537000 +trainer/Policy Loss -19.2659 +trainer/Log Pis Mean 24.664 +trainer/Log Pis Std 12.9605 +trainer/Log Pis Max 68.2256 +trainer/Log Pis Min -13.2884 +trainer/policy/mean Mean -0.0349693 +trainer/policy/mean Std 0.909543 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76629 +trainer/policy/normal/std Std 0.60984 +trainer/policy/normal/std Max 5.09561 +trainer/policy/normal/std Min 0.310659 +trainer/policy/normal/log_std Mean 0.982638 +trainer/policy/normal/log_std Std 0.297399 +trainer/policy/normal/log_std Max 1.62838 +trainer/policy/normal/log_std Min -1.16906 +eval/num steps total 535293 +eval/num paths total 537 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.157485 +eval/Actions Std 0.912508 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6966 +time/logging (s) 0.00365743 +time/sampling batch (s) 0.528995 +time/saving (s) 0.00345143 +time/training (s) 6.98109 +time/epoch (s) 10.2138 +time/total (s) 5579.55 +Epoch -464 +---------------------------------- --------------- +2022-05-10 14:43:53.385733 PDT | [1] Epoch -463 finished +---------------------------------- --------------- +epoch -463 +replay_buffer/size 999033 +trainer/num train calls 538000 +trainer/Policy Loss -19.3433 +trainer/Log Pis Mean 24.0437 +trainer/Log Pis Std 13.3181 +trainer/Log Pis Max 62.7748 +trainer/Log Pis Min -11.5092 +trainer/policy/mean Mean -0.0356089 +trainer/policy/mean Std 0.905687 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83044 +trainer/policy/normal/std Std 0.644568 +trainer/policy/normal/std Max 5.77278 +trainer/policy/normal/std Min 0.300799 +trainer/policy/normal/log_std Mean 1.00112 +trainer/policy/normal/log_std Std 0.321026 +trainer/policy/normal/log_std Max 1.75315 +trainer/policy/normal/log_std Min -1.20131 +eval/num steps total 536293 +eval/num paths total 538 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0180791 +eval/Actions Std 0.975105 +eval/Actions Max 0.999982 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51573 +time/logging (s) 0.00411606 +time/sampling batch (s) 0.53065 +time/saving (s) 0.00387117 +time/training (s) 7.55894 +time/epoch (s) 10.6133 +time/total (s) 5590.16 +Epoch -463 +---------------------------------- --------------- +2022-05-10 14:44:02.875358 PDT | [1] Epoch -462 finished +---------------------------------- --------------- +epoch -462 +replay_buffer/size 999033 +trainer/num train calls 539000 +trainer/Policy Loss -20.9039 +trainer/Log Pis Mean 24.8372 +trainer/Log Pis Std 13.2214 +trainer/Log Pis Max 69.1796 +trainer/Log Pis Min -6.4876 +trainer/policy/mean Mean -0.0465615 +trainer/policy/mean Std 0.901069 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7651 +trainer/policy/normal/std Std 0.637696 +trainer/policy/normal/std Max 5.66093 +trainer/policy/normal/std Min 0.278357 +trainer/policy/normal/log_std Mean 0.979451 +trainer/policy/normal/log_std Std 0.307417 +trainer/policy/normal/log_std Max 1.73359 +trainer/policy/normal/log_std Min -1.27885 +eval/num steps total 537293 +eval/num paths total 539 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0267925 +eval/Actions Std 0.901137 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44935 +time/logging (s) 0.00386326 +time/sampling batch (s) 0.281565 +time/saving (s) 0.00371696 +time/training (s) 6.72747 +time/epoch (s) 9.46596 +time/total (s) 5599.63 +Epoch -462 +---------------------------------- --------------- +2022-05-10 14:44:13.623215 PDT | [1] Epoch -461 finished +---------------------------------- --------------- +epoch -461 +replay_buffer/size 999033 +trainer/num train calls 540000 +trainer/Policy Loss -19.5636 +trainer/Log Pis Mean 24.608 +trainer/Log Pis Std 12.8663 +trainer/Log Pis Max 73.3293 +trainer/Log Pis Min -5.30204 +trainer/policy/mean Mean -0.0483531 +trainer/policy/mean Std 0.906166 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80907 +trainer/policy/normal/std Std 0.631136 +trainer/policy/normal/std Max 5.41658 +trainer/policy/normal/std Min 0.315065 +trainer/policy/normal/log_std Mean 0.997145 +trainer/policy/normal/log_std Std 0.299406 +trainer/policy/normal/log_std Max 1.68946 +trainer/policy/normal/log_std Min -1.15498 +eval/num steps total 538293 +eval/num paths total 540 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.349183 +eval/Actions Std 0.868961 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30829 +time/logging (s) 0.00369577 +time/sampling batch (s) 0.279109 +time/saving (s) 0.00351662 +time/training (s) 8.13019 +time/epoch (s) 10.7248 +time/total (s) 5610.36 +Epoch -461 +---------------------------------- --------------- +2022-05-10 14:44:24.266026 PDT | [1] Epoch -460 finished +---------------------------------- --------------- +epoch -460 +replay_buffer/size 999033 +trainer/num train calls 541000 +trainer/Policy Loss -18.7853 +trainer/Log Pis Mean 24.1962 +trainer/Log Pis Std 12.9838 +trainer/Log Pis Max 70.0336 +trainer/Log Pis Min -6.73352 +trainer/policy/mean Mean -0.0343314 +trainer/policy/mean Std 0.903255 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.77049 +trainer/policy/normal/std Std 0.633529 +trainer/policy/normal/std Max 5.44722 +trainer/policy/normal/std Min 0.275684 +trainer/policy/normal/log_std Mean 0.980448 +trainer/policy/normal/log_std Std 0.315784 +trainer/policy/normal/log_std Max 1.69511 +trainer/policy/normal/log_std Min -1.2885 +eval/num steps total 539293 +eval/num paths total 541 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00904135 +eval/Actions Std 0.911807 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50441 +time/logging (s) 0.00382061 +time/sampling batch (s) 0.279335 +time/saving (s) 0.00341043 +time/training (s) 7.82915 +time/epoch (s) 10.6201 +time/total (s) 5620.98 +Epoch -460 +---------------------------------- --------------- +2022-05-10 14:44:34.948189 PDT | [1] Epoch -459 finished +---------------------------------- --------------- +epoch -459 +replay_buffer/size 999033 +trainer/num train calls 542000 +trainer/Policy Loss -20.0565 +trainer/Log Pis Mean 24.8604 +trainer/Log Pis Std 13.2521 +trainer/Log Pis Max 68.9219 +trainer/Log Pis Min -4.27584 +trainer/policy/mean Mean -0.0485133 +trainer/policy/mean Std 0.908 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85696 +trainer/policy/normal/std Std 0.657174 +trainer/policy/normal/std Max 6.2978 +trainer/policy/normal/std Min 0.315404 +trainer/policy/normal/log_std Mean 1.01176 +trainer/policy/normal/log_std Std 0.31123 +trainer/policy/normal/log_std Max 1.8402 +trainer/policy/normal/log_std Min -1.1539 +eval/num steps total 540293 +eval/num paths total 542 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104189 +eval/Actions Std 0.914785 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71912 +time/logging (s) 0.00391071 +time/sampling batch (s) 0.280244 +time/saving (s) 0.00346505 +time/training (s) 7.65279 +time/epoch (s) 10.6595 +time/total (s) 5631.65 +Epoch -459 +---------------------------------- --------------- +2022-05-10 14:44:45.749060 PDT | [1] Epoch -458 finished +---------------------------------- --------------- +epoch -458 +replay_buffer/size 999033 +trainer/num train calls 543000 +trainer/Policy Loss -20.9951 +trainer/Log Pis Mean 23.836 +trainer/Log Pis Std 13.5428 +trainer/Log Pis Max 66.3874 +trainer/Log Pis Min -8.0823 +trainer/policy/mean Mean -0.0372376 +trainer/policy/mean Std 0.908945 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79901 +trainer/policy/normal/std Std 0.632298 +trainer/policy/normal/std Max 6.89201 +trainer/policy/normal/std Min 0.289776 +trainer/policy/normal/log_std Mean 0.99382 +trainer/policy/normal/log_std Std 0.296464 +trainer/policy/normal/log_std Max 1.93036 +trainer/policy/normal/log_std Min -1.23865 +eval/num steps total 541293 +eval/num paths total 543 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00465911 +eval/Actions Std 0.910686 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72203 +time/logging (s) 0.0041991 +time/sampling batch (s) 0.280108 +time/saving (s) 0.00385152 +time/training (s) 7.76808 +time/epoch (s) 10.7783 +time/total (s) 5642.43 +Epoch -458 +---------------------------------- --------------- +2022-05-10 14:44:56.256696 PDT | [1] Epoch -457 finished +---------------------------------- --------------- +epoch -457 +replay_buffer/size 999033 +trainer/num train calls 544000 +trainer/Policy Loss -19.755 +trainer/Log Pis Mean 23.7927 +trainer/Log Pis Std 13.0246 +trainer/Log Pis Max 65.8321 +trainer/Log Pis Min -8.76373 +trainer/policy/mean Mean -0.0230047 +trainer/policy/mean Std 0.902398 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.76812 +trainer/policy/normal/std Std 0.645901 +trainer/policy/normal/std Max 5.46122 +trainer/policy/normal/std Min 0.282083 +trainer/policy/normal/log_std Mean 0.978873 +trainer/policy/normal/log_std Std 0.315982 +trainer/policy/normal/log_std Max 1.69767 +trainer/policy/normal/log_std Min -1.26555 +eval/num steps total 542293 +eval/num paths total 544 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.071093 +eval/Actions Std 0.901775 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41593 +time/logging (s) 0.00389472 +time/sampling batch (s) 0.52905 +time/saving (s) 0.00369527 +time/training (s) 7.53156 +time/epoch (s) 10.4841 +time/total (s) 5652.92 +Epoch -457 +---------------------------------- --------------- +2022-05-10 14:45:06.567141 PDT | [1] Epoch -456 finished +---------------------------------- --------------- +epoch -456 +replay_buffer/size 999033 +trainer/num train calls 545000 +trainer/Policy Loss -20.1396 +trainer/Log Pis Mean 23.5739 +trainer/Log Pis Std 12.6916 +trainer/Log Pis Max 68.668 +trainer/Log Pis Min -9.80859 +trainer/policy/mean Mean -0.059164 +trainer/policy/mean Std 0.905669 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79553 +trainer/policy/normal/std Std 0.626311 +trainer/policy/normal/std Max 5.78852 +trainer/policy/normal/std Min 0.33141 +trainer/policy/normal/log_std Mean 0.990644 +trainer/policy/normal/log_std Std 0.311819 +trainer/policy/normal/log_std Max 1.75588 +trainer/policy/normal/log_std Min -1.1044 +eval/num steps total 543293 +eval/num paths total 545 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.462298 +eval/Actions Std 0.804852 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48018 +time/logging (s) 0.00384072 +time/sampling batch (s) 0.28011 +time/saving (s) 0.00345016 +time/training (s) 7.51973 +time/epoch (s) 10.2873 +time/total (s) 5663.21 +Epoch -456 +---------------------------------- --------------- +2022-05-10 14:45:16.242349 PDT | [1] Epoch -455 finished +---------------------------------- --------------- +epoch -455 +replay_buffer/size 999033 +trainer/num train calls 546000 +trainer/Policy Loss -20.5446 +trainer/Log Pis Mean 24.2931 +trainer/Log Pis Std 13.117 +trainer/Log Pis Max 68.7498 +trainer/Log Pis Min -7.18198 +trainer/policy/mean Mean -0.041755 +trainer/policy/mean Std 0.911077 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79246 +trainer/policy/normal/std Std 0.629036 +trainer/policy/normal/std Max 5.73063 +trainer/policy/normal/std Min 0.314727 +trainer/policy/normal/log_std Mean 0.990443 +trainer/policy/normal/log_std Std 0.304969 +trainer/policy/normal/log_std Max 1.74582 +trainer/policy/normal/log_std Min -1.15605 +eval/num steps total 544293 +eval/num paths total 546 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.124255 +eval/Actions Std 0.905145 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80325 +time/logging (s) 0.00390642 +time/sampling batch (s) 0.280077 +time/saving (s) 0.00344829 +time/training (s) 6.56155 +time/epoch (s) 9.65223 +time/total (s) 5672.86 +Epoch -455 +---------------------------------- --------------- +2022-05-10 14:45:25.956987 PDT | [1] Epoch -454 finished +---------------------------------- --------------- +epoch -454 +replay_buffer/size 999033 +trainer/num train calls 547000 +trainer/Policy Loss -19.9606 +trainer/Log Pis Mean 24.4603 +trainer/Log Pis Std 13.7845 +trainer/Log Pis Max 86.6732 +trainer/Log Pis Min -6.00923 +trainer/policy/mean Mean -0.0310176 +trainer/policy/mean Std 0.911257 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85323 +trainer/policy/normal/std Std 0.632379 +trainer/policy/normal/std Max 5.68932 +trainer/policy/normal/std Min 0.316347 +trainer/policy/normal/log_std Mean 1.01476 +trainer/policy/normal/log_std Std 0.289063 +trainer/policy/normal/log_std Max 1.73859 +trainer/policy/normal/log_std Min -1.15091 +eval/num steps total 545293 +eval/num paths total 547 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118904 +eval/Actions Std 0.899433 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65897 +time/logging (s) 0.00383267 +time/sampling batch (s) 0.280045 +time/saving (s) 0.00350194 +time/training (s) 6.74522 +time/epoch (s) 9.69157 +time/total (s) 5682.56 +Epoch -454 +---------------------------------- --------------- +2022-05-10 14:45:36.135132 PDT | [1] Epoch -453 finished +---------------------------------- --------------- +epoch -453 +replay_buffer/size 999033 +trainer/num train calls 548000 +trainer/Policy Loss -19.6578 +trainer/Log Pis Mean 24.5567 +trainer/Log Pis Std 12.597 +trainer/Log Pis Max 62.3533 +trainer/Log Pis Min -5.34784 +trainer/policy/mean Mean -0.0443484 +trainer/policy/mean Std 0.906267 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83326 +trainer/policy/normal/std Std 0.649606 +trainer/policy/normal/std Max 7.07203 +trainer/policy/normal/std Min 0.341247 +trainer/policy/normal/log_std Mean 1.00374 +trainer/policy/normal/log_std Std 0.309492 +trainer/policy/normal/log_std Max 1.95615 +trainer/policy/normal/log_std Min -1.07515 +eval/num steps total 546293 +eval/num paths total 548 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0922818 +eval/Actions Std 0.908067 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31183 +time/logging (s) 0.00378543 +time/sampling batch (s) 0.280743 +time/saving (s) 0.00339916 +time/training (s) 7.55522 +time/epoch (s) 10.155 +time/total (s) 5692.71 +Epoch -453 +---------------------------------- --------------- +2022-05-10 14:45:46.092464 PDT | [1] Epoch -452 finished +---------------------------------- --------------- +epoch -452 +replay_buffer/size 999033 +trainer/num train calls 549000 +trainer/Policy Loss -18.692 +trainer/Log Pis Mean 24.7377 +trainer/Log Pis Std 13.0547 +trainer/Log Pis Max 78.3433 +trainer/Log Pis Min -8.08928 +trainer/policy/mean Mean -0.0424271 +trainer/policy/mean Std 0.904496 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87234 +trainer/policy/normal/std Std 0.657835 +trainer/policy/normal/std Max 5.74766 +trainer/policy/normal/std Min 0.328351 +trainer/policy/normal/log_std Mean 1.01752 +trainer/policy/normal/log_std Std 0.309588 +trainer/policy/normal/log_std Max 1.74879 +trainer/policy/normal/log_std Min -1.11367 +eval/num steps total 547293 +eval/num paths total 549 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.233419 +eval/Actions Std 0.903191 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35752 +time/logging (s) 0.00407593 +time/sampling batch (s) 0.291586 +time/saving (s) 0.00409236 +time/training (s) 7.27726 +time/epoch (s) 9.93453 +time/total (s) 5702.65 +Epoch -452 +---------------------------------- --------------- +2022-05-10 14:45:56.904411 PDT | [1] Epoch -451 finished +---------------------------------- --------------- +epoch -451 +replay_buffer/size 999033 +trainer/num train calls 550000 +trainer/Policy Loss -19.3932 +trainer/Log Pis Mean 24.4713 +trainer/Log Pis Std 13.4135 +trainer/Log Pis Max 65.48 +trainer/Log Pis Min -5.08491 +trainer/policy/mean Mean -0.0385894 +trainer/policy/mean Std 0.907197 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7942 +trainer/policy/normal/std Std 0.64498 +trainer/policy/normal/std Max 7.383 +trainer/policy/normal/std Min 0.316994 +trainer/policy/normal/log_std Mean 0.989356 +trainer/policy/normal/log_std Std 0.311565 +trainer/policy/normal/log_std Max 1.99918 +trainer/policy/normal/log_std Min -1.14887 +eval/num steps total 548293 +eval/num paths total 550 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.464168 +eval/Actions Std 0.871958 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.90037 +time/logging (s) 0.00379829 +time/sampling batch (s) 0.30714 +time/saving (s) 0.00354387 +time/training (s) 7.57328 +time/epoch (s) 10.7881 +time/total (s) 5713.44 +Epoch -451 +---------------------------------- --------------- +2022-05-10 14:46:07.546307 PDT | [1] Epoch -450 finished +---------------------------------- --------------- +epoch -450 +replay_buffer/size 999033 +trainer/num train calls 551000 +trainer/Policy Loss -20.7739 +trainer/Log Pis Mean 23.6559 +trainer/Log Pis Std 13.1407 +trainer/Log Pis Max 77.1217 +trainer/Log Pis Min -7.00071 +trainer/policy/mean Mean -0.0322451 +trainer/policy/mean Std 0.906029 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81625 +trainer/policy/normal/std Std 0.649699 +trainer/policy/normal/std Max 5.92755 +trainer/policy/normal/std Min 0.277332 +trainer/policy/normal/log_std Mean 0.997763 +trainer/policy/normal/log_std Std 0.309078 +trainer/policy/normal/log_std Max 1.77961 +trainer/policy/normal/log_std Min -1.28254 +eval/num steps total 549293 +eval/num paths total 551 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0769047 +eval/Actions Std 0.906999 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68627 +time/logging (s) 0.00373623 +time/sampling batch (s) 0.283823 +time/saving (s) 0.00342152 +time/training (s) 7.64138 +time/epoch (s) 10.6186 +time/total (s) 5724.06 +Epoch -450 +---------------------------------- --------------- +2022-05-10 14:46:17.338820 PDT | [1] Epoch -449 finished +---------------------------------- --------------- +epoch -449 +replay_buffer/size 999033 +trainer/num train calls 552000 +trainer/Policy Loss -19.9045 +trainer/Log Pis Mean 25.4872 +trainer/Log Pis Std 12.9225 +trainer/Log Pis Max 59.4419 +trainer/Log Pis Min -7.95553 +trainer/policy/mean Mean -0.0347622 +trainer/policy/mean Std 0.90518 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80002 +trainer/policy/normal/std Std 0.656627 +trainer/policy/normal/std Max 5.13537 +trainer/policy/normal/std Min 0.339919 +trainer/policy/normal/log_std Mean 0.989094 +trainer/policy/normal/log_std Std 0.323345 +trainer/policy/normal/log_std Max 1.63615 +trainer/policy/normal/log_std Min -1.07905 +eval/num steps total 550293 +eval/num paths total 552 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0701598 +eval/Actions Std 0.912332 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61749 +time/logging (s) 0.00375993 +time/sampling batch (s) 0.781177 +time/saving (s) 0.00342303 +time/training (s) 6.36364 +time/epoch (s) 9.7695 +time/total (s) 5733.84 +Epoch -449 +---------------------------------- --------------- +2022-05-10 14:46:27.223469 PDT | [1] Epoch -448 finished +---------------------------------- --------------- +epoch -448 +replay_buffer/size 999033 +trainer/num train calls 553000 +trainer/Policy Loss -21.939 +trainer/Log Pis Mean 24.274 +trainer/Log Pis Std 12.9094 +trainer/Log Pis Max 65.588 +trainer/Log Pis Min -5.57795 +trainer/policy/mean Mean -0.0583675 +trainer/policy/mean Std 0.905811 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82469 +trainer/policy/normal/std Std 0.655817 +trainer/policy/normal/std Max 5.42197 +trainer/policy/normal/std Min 0.309477 +trainer/policy/normal/log_std Mean 0.998202 +trainer/policy/normal/log_std Std 0.322396 +trainer/policy/normal/log_std Max 1.69046 +trainer/policy/normal/log_std Min -1.17287 +eval/num steps total 551293 +eval/num paths total 553 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.214502 +eval/Actions Std 0.850479 +eval/Actions Max 0.999986 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51325 +time/logging (s) 0.00382255 +time/sampling batch (s) 0.525493 +time/saving (s) 0.00342414 +time/training (s) 6.8161 +time/epoch (s) 9.86209 +time/total (s) 5743.7 +Epoch -448 +---------------------------------- --------------- +2022-05-10 14:46:37.339540 PDT | [1] Epoch -447 finished +---------------------------------- --------------- +epoch -447 +replay_buffer/size 999033 +trainer/num train calls 554000 +trainer/Policy Loss -20.7032 +trainer/Log Pis Mean 23.6617 +trainer/Log Pis Std 13.9342 +trainer/Log Pis Max 77.9218 +trainer/Log Pis Min -6.29843 +trainer/policy/mean Mean -0.0325326 +trainer/policy/mean Std 0.905689 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.85469 +trainer/policy/normal/std Std 0.643431 +trainer/policy/normal/std Max 5.30671 +trainer/policy/normal/std Min 0.324064 +trainer/policy/normal/log_std Mean 1.01324 +trainer/policy/normal/log_std Std 0.298905 +trainer/policy/normal/log_std Max 1.66897 +trainer/policy/normal/log_std Min -1.12681 +eval/num steps total 552293 +eval/num paths total 554 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.277854 +eval/Actions Std 0.886107 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70955 +time/logging (s) 0.00404438 +time/sampling batch (s) 0.526311 +time/saving (s) 0.00381401 +time/training (s) 6.84988 +time/epoch (s) 10.0936 +time/total (s) 5753.8 +Epoch -447 +---------------------------------- --------------- +2022-05-10 14:46:48.231972 PDT | [1] Epoch -446 finished +---------------------------------- --------------- +epoch -446 +replay_buffer/size 999033 +trainer/num train calls 555000 +trainer/Policy Loss -19.0603 +trainer/Log Pis Mean 24.4203 +trainer/Log Pis Std 13.3713 +trainer/Log Pis Max 76.6949 +trainer/Log Pis Min -8.84631 +trainer/policy/mean Mean -0.0268251 +trainer/policy/mean Std 0.903375 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8058 +trainer/policy/normal/std Std 0.647012 +trainer/policy/normal/std Max 5.70365 +trainer/policy/normal/std Min 0.302004 +trainer/policy/normal/log_std Mean 0.994351 +trainer/policy/normal/log_std Std 0.305547 +trainer/policy/normal/log_std Max 1.74111 +trainer/policy/normal/log_std Min -1.19732 +eval/num steps total 553293 +eval/num paths total 555 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0580029 +eval/Actions Std 0.894301 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66612 +time/logging (s) 0.00378652 +time/sampling batch (s) 0.279516 +time/saving (s) 0.00367848 +time/training (s) 7.91538 +time/epoch (s) 10.8685 +time/total (s) 5764.67 +Epoch -446 +---------------------------------- --------------- +2022-05-10 14:46:58.107474 PDT | [1] Epoch -445 finished +---------------------------------- --------------- +epoch -445 +replay_buffer/size 999033 +trainer/num train calls 556000 +trainer/Policy Loss -20.5266 +trainer/Log Pis Mean 24.873 +trainer/Log Pis Std 13.6479 +trainer/Log Pis Max 74.0696 +trainer/Log Pis Min -11.44 +trainer/policy/mean Mean -0.0330645 +trainer/policy/mean Std 0.903959 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78303 +trainer/policy/normal/std Std 0.678762 +trainer/policy/normal/std Max 5.96034 +trainer/policy/normal/std Min 0.231026 +trainer/policy/normal/log_std Mean 0.979005 +trainer/policy/normal/log_std Std 0.340752 +trainer/policy/normal/log_std Max 1.78513 +trainer/policy/normal/log_std Min -1.46522 +eval/num steps total 554293 +eval/num paths total 556 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118851 +eval/Actions Std 0.887468 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53691 +time/logging (s) 0.00370048 +time/sampling batch (s) 0.278716 +time/saving (s) 0.00348188 +time/training (s) 7.02973 +time/epoch (s) 9.85254 +time/total (s) 5774.53 +Epoch -445 +---------------------------------- --------------- +2022-05-10 14:47:08.548616 PDT | [1] Epoch -444 finished +---------------------------------- --------------- +epoch -444 +replay_buffer/size 999033 +trainer/num train calls 557000 +trainer/Policy Loss -18.4937 +trainer/Log Pis Mean 24.6933 +trainer/Log Pis Std 13.6549 +trainer/Log Pis Max 70.1668 +trainer/Log Pis Min -9.85754 +trainer/policy/mean Mean -0.0351205 +trainer/policy/mean Std 0.909457 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80117 +trainer/policy/normal/std Std 0.656967 +trainer/policy/normal/std Max 5.90659 +trainer/policy/normal/std Min 0.241071 +trainer/policy/normal/log_std Mean 0.989326 +trainer/policy/normal/log_std Std 0.326277 +trainer/policy/normal/log_std Max 1.77607 +trainer/policy/normal/log_std Min -1.42267 +eval/num steps total 555133 +eval/num paths total 557 +eval/path length Mean 840 +eval/path length Std 0 +eval/path length Max 840 +eval/path length Min 840 +eval/Rewards Mean 0.00119048 +eval/Rewards Std 0.0344827 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.031104 +eval/Actions Std 0.907934 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.61324 +time/logging (s) 0.00329653 +time/sampling batch (s) 0.528753 +time/saving (s) 0.00345477 +time/training (s) 7.26906 +time/epoch (s) 10.4178 +time/total (s) 5784.95 +Epoch -444 +---------------------------------- --------------- +2022-05-10 14:47:18.877147 PDT | [1] Epoch -443 finished +---------------------------------- --------------- +epoch -443 +replay_buffer/size 999033 +trainer/num train calls 558000 +trainer/Policy Loss -19.1006 +trainer/Log Pis Mean 24.3847 +trainer/Log Pis Std 13.0342 +trainer/Log Pis Max 72.1316 +trainer/Log Pis Min -8.20216 +trainer/policy/mean Mean -0.0309962 +trainer/policy/mean Std 0.902558 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82512 +trainer/policy/normal/std Std 0.667902 +trainer/policy/normal/std Max 7.31304 +trainer/policy/normal/std Min 0.31068 +trainer/policy/normal/log_std Mean 0.999241 +trainer/policy/normal/log_std Std 0.31446 +trainer/policy/normal/log_std Max 1.98966 +trainer/policy/normal/log_std Min -1.16899 +eval/num steps total 556133 +eval/num paths total 558 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0261678 +eval/Actions Std 0.910239 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66041 +time/logging (s) 0.00369959 +time/sampling batch (s) 0.529593 +time/saving (s) 0.0035068 +time/training (s) 7.10878 +time/epoch (s) 10.306 +time/total (s) 5795.26 +Epoch -443 +---------------------------------- --------------- +2022-05-10 14:47:28.928154 PDT | [1] Epoch -442 finished +---------------------------------- --------------- +epoch -442 +replay_buffer/size 999033 +trainer/num train calls 559000 +trainer/Policy Loss -19.4578 +trainer/Log Pis Mean 24.2488 +trainer/Log Pis Std 13.3821 +trainer/Log Pis Max 69.0162 +trainer/Log Pis Min -8.92626 +trainer/policy/mean Mean -0.0340901 +trainer/policy/mean Std 0.903961 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.78325 +trainer/policy/normal/std Std 0.657919 +trainer/policy/normal/std Max 6.42286 +trainer/policy/normal/std Min 0.288063 +trainer/policy/normal/log_std Mean 0.983644 +trainer/policy/normal/log_std Std 0.318544 +trainer/policy/normal/log_std Max 1.85986 +trainer/policy/normal/log_std Min -1.24458 +eval/num steps total 557133 +eval/num paths total 559 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.485263 +eval/Actions Std 0.86715 +eval/Actions Max 0.999954 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63888 +time/logging (s) 0.00412808 +time/sampling batch (s) 0.279632 +time/saving (s) 0.0038366 +time/training (s) 7.10201 +time/epoch (s) 10.0285 +time/total (s) 5805.29 +Epoch -442 +---------------------------------- --------------- +2022-05-10 14:47:37.985236 PDT | [1] Epoch -441 finished +---------------------------------- --------------- +epoch -441 +replay_buffer/size 999033 +trainer/num train calls 560000 +trainer/Policy Loss -19.9394 +trainer/Log Pis Mean 25.0971 +trainer/Log Pis Std 13.3105 +trainer/Log Pis Max 64.7383 +trainer/Log Pis Min -11.738 +trainer/policy/mean Mean -0.0362737 +trainer/policy/mean Std 0.906148 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78948 +trainer/policy/normal/std Std 0.657114 +trainer/policy/normal/std Max 6.55329 +trainer/policy/normal/std Min 0.288811 +trainer/policy/normal/log_std Mean 0.984917 +trainer/policy/normal/log_std Std 0.325182 +trainer/policy/normal/log_std Max 1.87997 +trainer/policy/normal/log_std Min -1.24198 +eval/num steps total 558133 +eval/num paths total 560 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0661804 +eval/Actions Std 0.918732 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53543 +time/logging (s) 0.00394935 +time/sampling batch (s) 0.281397 +time/saving (s) 0.00382375 +time/training (s) 6.20859 +time/epoch (s) 9.03319 +time/total (s) 5814.32 +Epoch -441 +---------------------------------- --------------- +2022-05-10 14:47:48.437440 PDT | [1] Epoch -440 finished +---------------------------------- --------------- +epoch -440 +replay_buffer/size 999033 +trainer/num train calls 561000 +trainer/Policy Loss -19.1858 +trainer/Log Pis Mean 23.4358 +trainer/Log Pis Std 13.4044 +trainer/Log Pis Max 72.8249 +trainer/Log Pis Min -10.6684 +trainer/policy/mean Mean -0.0185577 +trainer/policy/mean Std 0.90498 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79262 +trainer/policy/normal/std Std 0.65982 +trainer/policy/normal/std Max 5.42746 +trainer/policy/normal/std Min 0.298582 +trainer/policy/normal/log_std Mean 0.986404 +trainer/policy/normal/log_std Std 0.322163 +trainer/policy/normal/log_std Max 1.69147 +trainer/policy/normal/log_std Min -1.20871 +eval/num steps total 559133 +eval/num paths total 561 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.163508 +eval/Actions Std 0.861993 +eval/Actions Max 0.999985 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46561 +time/logging (s) 0.00378789 +time/sampling batch (s) 0.290066 +time/saving (s) 0.0034853 +time/training (s) 7.66521 +time/epoch (s) 10.4282 +time/total (s) 5824.75 +Epoch -440 +---------------------------------- --------------- +2022-05-10 14:47:58.072038 PDT | [1] Epoch -439 finished +---------------------------------- --------------- +epoch -439 +replay_buffer/size 999033 +trainer/num train calls 562000 +trainer/Policy Loss -19.8982 +trainer/Log Pis Mean 24.4275 +trainer/Log Pis Std 13.8103 +trainer/Log Pis Max 67.9793 +trainer/Log Pis Min -4.12294 +trainer/policy/mean Mean -0.0365226 +trainer/policy/mean Std 0.907048 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83312 +trainer/policy/normal/std Std 0.657803 +trainer/policy/normal/std Max 6.16102 +trainer/policy/normal/std Min 0.24484 +trainer/policy/normal/log_std Mean 1.00048 +trainer/policy/normal/log_std Std 0.328676 +trainer/policy/normal/log_std Max 1.81824 +trainer/policy/normal/log_std Min -1.40715 +eval/num steps total 560133 +eval/num paths total 562 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0645047 +eval/Actions Std 0.899617 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66045 +time/logging (s) 0.0037192 +time/sampling batch (s) 0.282735 +time/saving (s) 0.00350029 +time/training (s) 6.66081 +time/epoch (s) 9.61121 +time/total (s) 5834.37 +Epoch -439 +---------------------------------- --------------- +2022-05-10 14:48:07.967766 PDT | [1] Epoch -438 finished +---------------------------------- --------------- +epoch -438 +replay_buffer/size 999033 +trainer/num train calls 563000 +trainer/Policy Loss -17.993 +trainer/Log Pis Mean 23.9087 +trainer/Log Pis Std 13.1293 +trainer/Log Pis Max 67.5854 +trainer/Log Pis Min -3.72817 +trainer/policy/mean Mean -0.0222777 +trainer/policy/mean Std 0.907596 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78048 +trainer/policy/normal/std Std 0.617967 +trainer/policy/normal/std Max 6.14671 +trainer/policy/normal/std Min 0.318973 +trainer/policy/normal/log_std Mean 0.988635 +trainer/policy/normal/log_std Std 0.290616 +trainer/policy/normal/log_std Max 1.81592 +trainer/policy/normal/log_std Min -1.14265 +eval/num steps total 561133 +eval/num paths total 563 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.484439 +eval/Actions Std 0.810107 +eval/Actions Max 0.999998 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75201 +time/logging (s) 0.00380556 +time/sampling batch (s) 0.282776 +time/saving (s) 0.00343432 +time/training (s) 6.83066 +time/epoch (s) 9.87268 +time/total (s) 5844.24 +Epoch -438 +---------------------------------- --------------- +2022-05-10 14:48:16.868722 PDT | [1] Epoch -437 finished +---------------------------------- --------------- +epoch -437 +replay_buffer/size 999033 +trainer/num train calls 564000 +trainer/Policy Loss -20.1303 +trainer/Log Pis Mean 23.3047 +trainer/Log Pis Std 12.3969 +trainer/Log Pis Max 61.6671 +trainer/Log Pis Min -7.49758 +trainer/policy/mean Mean -0.0418874 +trainer/policy/mean Std 0.902939 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81383 +trainer/policy/normal/std Std 0.634751 +trainer/policy/normal/std Max 5.01437 +trainer/policy/normal/std Min 0.279939 +trainer/policy/normal/log_std Mean 0.997872 +trainer/policy/normal/log_std Std 0.305507 +trainer/policy/normal/log_std Max 1.61231 +trainer/policy/normal/log_std Min -1.27318 +eval/num steps total 562133 +eval/num paths total 564 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0314494 +eval/Actions Std 0.898527 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47833 +time/logging (s) 0.00370637 +time/sampling batch (s) 0.282471 +time/saving (s) 0.00349607 +time/training (s) 6.10957 +time/epoch (s) 8.87758 +time/total (s) 5853.12 +Epoch -437 +---------------------------------- --------------- +2022-05-10 14:48:27.405546 PDT | [1] Epoch -436 finished +---------------------------------- --------------- +epoch -436 +replay_buffer/size 999033 +trainer/num train calls 565000 +trainer/Policy Loss -19.8627 +trainer/Log Pis Mean 23.2532 +trainer/Log Pis Std 12.9603 +trainer/Log Pis Max 63.3787 +trainer/Log Pis Min -6.24012 +trainer/policy/mean Mean -0.0288186 +trainer/policy/mean Std 0.90632 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78153 +trainer/policy/normal/std Std 0.638735 +trainer/policy/normal/std Max 5.75403 +trainer/policy/normal/std Min 0.323194 +trainer/policy/normal/log_std Mean 0.985987 +trainer/policy/normal/log_std Std 0.305371 +trainer/policy/normal/log_std Max 1.7499 +trainer/policy/normal/log_std Min -1.1295 +eval/num steps total 563133 +eval/num paths total 565 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.10913 +eval/Actions Std 0.906532 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52529 +time/logging (s) 0.00419408 +time/sampling batch (s) 0.282556 +time/saving (s) 0.00380144 +time/training (s) 7.6981 +time/epoch (s) 10.5139 +time/total (s) 5863.64 +Epoch -436 +---------------------------------- --------------- +2022-05-10 14:48:38.636433 PDT | [1] Epoch -435 finished +---------------------------------- --------------- +epoch -435 +replay_buffer/size 999033 +trainer/num train calls 566000 +trainer/Policy Loss -19.6816 +trainer/Log Pis Mean 23.6584 +trainer/Log Pis Std 13.1615 +trainer/Log Pis Max 63.1275 +trainer/Log Pis Min -8.89143 +trainer/policy/mean Mean -0.0413187 +trainer/policy/mean Std 0.905765 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79841 +trainer/policy/normal/std Std 0.658069 +trainer/policy/normal/std Max 5.29189 +trainer/policy/normal/std Min 0.236321 +trainer/policy/normal/log_std Mean 0.988638 +trainer/policy/normal/log_std Std 0.322491 +trainer/policy/normal/log_std Max 1.66618 +trainer/policy/normal/log_std Min -1.44257 +eval/num steps total 564133 +eval/num paths total 566 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.236967 +eval/Actions Std 0.897643 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81503 +time/logging (s) 0.00379183 +time/sampling batch (s) 0.280069 +time/saving (s) 0.00372862 +time/training (s) 8.10439 +time/epoch (s) 11.207 +time/total (s) 5874.85 +Epoch -435 +---------------------------------- --------------- +2022-05-10 14:48:49.160311 PDT | [1] Epoch -434 finished +---------------------------------- --------------- +epoch -434 +replay_buffer/size 999033 +trainer/num train calls 567000 +trainer/Policy Loss -18.8556 +trainer/Log Pis Mean 24.7299 +trainer/Log Pis Std 12.9984 +trainer/Log Pis Max 68.2313 +trainer/Log Pis Min -4.9034 +trainer/policy/mean Mean -0.0312278 +trainer/policy/mean Std 0.906635 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.80686 +trainer/policy/normal/std Std 0.645783 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.276369 +trainer/policy/normal/log_std Mean 0.994036 +trainer/policy/normal/log_std Std 0.311386 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.28602 +eval/num steps total 565133 +eval/num paths total 567 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.149734 +eval/Actions Std 0.933697 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59874 +time/logging (s) 0.00375287 +time/sampling batch (s) 0.529167 +time/saving (s) 0.00346296 +time/training (s) 7.3657 +time/epoch (s) 10.5008 +time/total (s) 5885.36 +Epoch -434 +---------------------------------- --------------- +2022-05-10 14:49:00.318666 PDT | [1] Epoch -433 finished +---------------------------------- --------------- +epoch -433 +replay_buffer/size 999033 +trainer/num train calls 568000 +trainer/Policy Loss -19.567 +trainer/Log Pis Mean 25.0696 +trainer/Log Pis Std 13.3745 +trainer/Log Pis Max 88.4977 +trainer/Log Pis Min -3.48617 +trainer/policy/mean Mean -0.0451356 +trainer/policy/mean Std 0.905392 +trainer/policy/mean Max 0.999966 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81322 +trainer/policy/normal/std Std 0.65313 +trainer/policy/normal/std Max 5.58365 +trainer/policy/normal/std Min 0.280933 +trainer/policy/normal/log_std Mean 0.994643 +trainer/policy/normal/log_std Std 0.320187 +trainer/policy/normal/log_std Max 1.71984 +trainer/policy/normal/log_std Min -1.26964 +eval/num steps total 566133 +eval/num paths total 568 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00804259 +eval/Actions Std 0.908773 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62812 +time/logging (s) 0.00370873 +time/sampling batch (s) 0.527725 +time/saving (s) 0.00339368 +time/training (s) 7.97256 +time/epoch (s) 11.1355 +time/total (s) 5896.49 +Epoch -433 +---------------------------------- --------------- +2022-05-10 14:49:10.840869 PDT | [1] Epoch -432 finished +---------------------------------- --------------- +epoch -432 +replay_buffer/size 999033 +trainer/num train calls 569000 +trainer/Policy Loss -19.5623 +trainer/Log Pis Mean 25.1607 +trainer/Log Pis Std 13.1336 +trainer/Log Pis Max 68.5165 +trainer/Log Pis Min -9.46022 +trainer/policy/mean Mean -0.0509246 +trainer/policy/mean Std 0.906281 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.88416 +trainer/policy/normal/std Std 0.650227 +trainer/policy/normal/std Max 5.99354 +trainer/policy/normal/std Min 0.346042 +trainer/policy/normal/log_std Mean 1.02425 +trainer/policy/normal/log_std Std 0.294644 +trainer/policy/normal/log_std Max 1.79068 +trainer/policy/normal/log_std Min -1.0612 +eval/num steps total 567133 +eval/num paths total 569 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0496652 +eval/Actions Std 0.912822 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59032 +time/logging (s) 0.00383135 +time/sampling batch (s) 0.277819 +time/saving (s) 0.00345028 +time/training (s) 7.62407 +time/epoch (s) 10.4995 +time/total (s) 5907 +Epoch -432 +---------------------------------- --------------- +2022-05-10 14:49:20.926963 PDT | [1] Epoch -431 finished +---------------------------------- --------------- +epoch -431 +replay_buffer/size 999033 +trainer/num train calls 570000 +trainer/Policy Loss -19.2241 +trainer/Log Pis Mean 23.7821 +trainer/Log Pis Std 13.3355 +trainer/Log Pis Max 64.1091 +trainer/Log Pis Min -7.02498 +trainer/policy/mean Mean -0.0226977 +trainer/policy/mean Std 0.907689 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999973 +trainer/policy/normal/std Mean 2.84467 +trainer/policy/normal/std Std 0.652098 +trainer/policy/normal/std Max 6.01286 +trainer/policy/normal/std Min 0.283647 +trainer/policy/normal/log_std Mean 1.00762 +trainer/policy/normal/log_std Std 0.310709 +trainer/policy/normal/log_std Max 1.7939 +trainer/policy/normal/log_std Min -1.26003 +eval/num steps total 568133 +eval/num paths total 570 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.332111 +eval/Actions Std 0.870553 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51509 +time/logging (s) 0.00436674 +time/sampling batch (s) 0.539357 +time/saving (s) 0.00422843 +time/training (s) 7.00058 +time/epoch (s) 10.0636 +time/total (s) 5917.06 +Epoch -431 +---------------------------------- --------------- +2022-05-10 14:49:32.658525 PDT | [1] Epoch -430 finished +---------------------------------- --------------- +epoch -430 +replay_buffer/size 999033 +trainer/num train calls 571000 +trainer/Policy Loss -20.8091 +trainer/Log Pis Mean 24.8019 +trainer/Log Pis Std 13.3499 +trainer/Log Pis Max 67.1957 +trainer/Log Pis Min -10.416 +trainer/policy/mean Mean -0.0430506 +trainer/policy/mean Std 0.907698 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82896 +trainer/policy/normal/std Std 0.649345 +trainer/policy/normal/std Max 5.13027 +trainer/policy/normal/std Min 0.205304 +trainer/policy/normal/log_std Mean 1.00091 +trainer/policy/normal/log_std Std 0.317946 +trainer/policy/normal/log_std Max 1.63516 +trainer/policy/normal/log_std Min -1.58326 +eval/num steps total 569133 +eval/num paths total 571 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0254208 +eval/Actions Std 0.902454 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74956 +time/logging (s) 0.0043017 +time/sampling batch (s) 0.306157 +time/saving (s) 0.00426264 +time/training (s) 8.64253 +time/epoch (s) 11.7068 +time/total (s) 5928.77 +Epoch -430 +---------------------------------- --------------- +2022-05-10 14:49:42.465796 PDT | [1] Epoch -429 finished +---------------------------------- --------------- +epoch -429 +replay_buffer/size 999033 +trainer/num train calls 572000 +trainer/Policy Loss -19.9608 +trainer/Log Pis Mean 24.0921 +trainer/Log Pis Std 13.4709 +trainer/Log Pis Max 79.9719 +trainer/Log Pis Min -6.2608 +trainer/policy/mean Mean -0.0308721 +trainer/policy/mean Std 0.904724 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83615 +trainer/policy/normal/std Std 0.653196 +trainer/policy/normal/std Max 6.32554 +trainer/policy/normal/std Min 0.328675 +trainer/policy/normal/log_std Mean 1.00534 +trainer/policy/normal/log_std Std 0.305629 +trainer/policy/normal/log_std Max 1.8446 +trainer/policy/normal/log_std Min -1.11269 +eval/num steps total 570133 +eval/num paths total 572 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0513228 +eval/Actions Std 0.877172 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50541 +time/logging (s) 0.0042571 +time/sampling batch (s) 0.287161 +time/saving (s) 0.00416434 +time/training (s) 6.98238 +time/epoch (s) 9.78337 +time/total (s) 5938.56 +Epoch -429 +---------------------------------- --------------- +2022-05-10 14:49:52.624000 PDT | [1] Epoch -428 finished +---------------------------------- --------------- +epoch -428 +replay_buffer/size 999033 +trainer/num train calls 573000 +trainer/Policy Loss -20.6384 +trainer/Log Pis Mean 24.6314 +trainer/Log Pis Std 13.5243 +trainer/Log Pis Max 64.7682 +trainer/Log Pis Min -4.92722 +trainer/policy/mean Mean -0.033404 +trainer/policy/mean Std 0.906038 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76915 +trainer/policy/normal/std Std 0.649163 +trainer/policy/normal/std Max 7.22228 +trainer/policy/normal/std Min 0.33273 +trainer/policy/normal/log_std Mean 0.980729 +trainer/policy/normal/log_std Std 0.306781 +trainer/policy/normal/log_std Max 1.97717 +trainer/policy/normal/log_std Min -1.10042 +eval/num steps total 571133 +eval/num paths total 573 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0773124 +eval/Actions Std 0.91143 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7048 +time/logging (s) 0.00421298 +time/sampling batch (s) 0.282697 +time/saving (s) 0.00409975 +time/training (s) 7.13838 +time/epoch (s) 10.1342 +time/total (s) 5948.7 +Epoch -428 +---------------------------------- --------------- +2022-05-10 14:50:02.965963 PDT | [1] Epoch -427 finished +---------------------------------- --------------- +epoch -427 +replay_buffer/size 999033 +trainer/num train calls 574000 +trainer/Policy Loss -18.436 +trainer/Log Pis Mean 24.7473 +trainer/Log Pis Std 12.6645 +trainer/Log Pis Max 65.6454 +trainer/Log Pis Min -5.38341 +trainer/policy/mean Mean -0.0383368 +trainer/policy/mean Std 0.905543 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81302 +trainer/policy/normal/std Std 0.656095 +trainer/policy/normal/std Max 6.15985 +trainer/policy/normal/std Min 0.315033 +trainer/policy/normal/log_std Mean 0.994505 +trainer/policy/normal/log_std Std 0.31985 +trainer/policy/normal/log_std Max 1.81805 +trainer/policy/normal/log_std Min -1.15508 +eval/num steps total 572133 +eval/num paths total 574 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0943756 +eval/Actions Std 0.915427 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62647 +time/logging (s) 0.00373195 +time/sampling batch (s) 0.5299 +time/saving (s) 0.00356049 +time/training (s) 7.15435 +time/epoch (s) 10.318 +time/total (s) 5959.02 +Epoch -427 +---------------------------------- --------------- +2022-05-10 14:50:12.626574 PDT | [1] Epoch -426 finished +---------------------------------- --------------- +epoch -426 +replay_buffer/size 999033 +trainer/num train calls 575000 +trainer/Policy Loss -19.0688 +trainer/Log Pis Mean 24.0034 +trainer/Log Pis Std 13.4184 +trainer/Log Pis Max 65.6058 +trainer/Log Pis Min -7.86053 +trainer/policy/mean Mean -0.038351 +trainer/policy/mean Std 0.909127 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81217 +trainer/policy/normal/std Std 0.626506 +trainer/policy/normal/std Max 5.96138 +trainer/policy/normal/std Min 0.245764 +trainer/policy/normal/log_std Mean 0.998919 +trainer/policy/normal/log_std Std 0.296563 +trainer/policy/normal/log_std Max 1.7853 +trainer/policy/normal/log_std Min -1.40338 +eval/num steps total 573133 +eval/num paths total 575 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.024877 +eval/Actions Std 0.908023 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47871 +time/logging (s) 0.00423849 +time/sampling batch (s) 0.278944 +time/saving (s) 0.00406908 +time/training (s) 6.87207 +time/epoch (s) 9.63803 +time/total (s) 5968.66 +Epoch -426 +---------------------------------- --------------- +2022-05-10 14:50:22.649791 PDT | [1] Epoch -425 finished +---------------------------------- --------------- +epoch -425 +replay_buffer/size 999033 +trainer/num train calls 576000 +trainer/Policy Loss -19.2202 +trainer/Log Pis Mean 23.8994 +trainer/Log Pis Std 13.4916 +trainer/Log Pis Max 63.3486 +trainer/Log Pis Min -6.30347 +trainer/policy/mean Mean -0.0286974 +trainer/policy/mean Std 0.90573 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.77247 +trainer/policy/normal/std Std 0.668449 +trainer/policy/normal/std Max 6.05895 +trainer/policy/normal/std Min 0.310383 +trainer/policy/normal/log_std Mean 0.976869 +trainer/policy/normal/log_std Std 0.332709 +trainer/policy/normal/log_std Max 1.80154 +trainer/policy/normal/log_std Min -1.16995 +eval/num steps total 574133 +eval/num paths total 576 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.376871 +eval/Actions Std 0.84455 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54084 +time/logging (s) 0.00384853 +time/sampling batch (s) 0.285683 +time/saving (s) 0.00342903 +time/training (s) 7.16517 +time/epoch (s) 9.99897 +time/total (s) 5978.66 +Epoch -425 +---------------------------------- --------------- +2022-05-10 14:50:32.351236 PDT | [1] Epoch -424 finished +---------------------------------- --------------- +epoch -424 +replay_buffer/size 999033 +trainer/num train calls 577000 +trainer/Policy Loss -19.3195 +trainer/Log Pis Mean 24.3165 +trainer/Log Pis Std 13.4466 +trainer/Log Pis Max 64.2223 +trainer/Log Pis Min -9.80189 +trainer/policy/mean Mean -0.0474005 +trainer/policy/mean Std 0.903917 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78402 +trainer/policy/normal/std Std 0.659005 +trainer/policy/normal/std Max 6.5835 +trainer/policy/normal/std Min 0.236071 +trainer/policy/normal/log_std Mean 0.982233 +trainer/policy/normal/log_std Std 0.329775 +trainer/policy/normal/log_std Max 1.88457 +trainer/policy/normal/log_std Min -1.44362 +eval/num steps total 575133 +eval/num paths total 577 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0851439 +eval/Actions Std 0.901439 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52131 +time/logging (s) 0.00429761 +time/sampling batch (s) 0.28653 +time/saving (s) 0.00427066 +time/training (s) 6.86208 +time/epoch (s) 9.67849 +time/total (s) 5988.34 +Epoch -424 +---------------------------------- --------------- +2022-05-10 14:50:43.145430 PDT | [1] Epoch -423 finished +---------------------------------- --------------- +epoch -423 +replay_buffer/size 999033 +trainer/num train calls 578000 +trainer/Policy Loss -20.026 +trainer/Log Pis Mean 24.5748 +trainer/Log Pis Std 13.7201 +trainer/Log Pis Max 61.4366 +trainer/Log Pis Min -5.24626 +trainer/policy/mean Mean -0.0268785 +trainer/policy/mean Std 0.904068 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82009 +trainer/policy/normal/std Std 0.623705 +trainer/policy/normal/std Max 6.77513 +trainer/policy/normal/std Min 0.266582 +trainer/policy/normal/log_std Mean 1.00211 +trainer/policy/normal/log_std Std 0.296825 +trainer/policy/normal/log_std Max 1.91326 +trainer/policy/normal/log_std Min -1.32207 +eval/num steps total 576133 +eval/num paths total 578 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.276173 +eval/Actions Std 0.908416 +eval/Actions Max 0.999995 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.91684 +time/logging (s) 0.00413442 +time/sampling batch (s) 0.784679 +time/saving (s) 0.00407356 +time/training (s) 7.06044 +time/epoch (s) 10.7702 +time/total (s) 5999.12 +Epoch -423 +---------------------------------- --------------- +2022-05-10 14:50:53.063730 PDT | [1] Epoch -422 finished +---------------------------------- --------------- +epoch -422 +replay_buffer/size 999033 +trainer/num train calls 579000 +trainer/Policy Loss -18.3202 +trainer/Log Pis Mean 24.4754 +trainer/Log Pis Std 13.2741 +trainer/Log Pis Max 71.6787 +trainer/Log Pis Min -9.27872 +trainer/policy/mean Mean -0.0573565 +trainer/policy/mean Std 0.903469 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79394 +trainer/policy/normal/std Std 0.652952 +trainer/policy/normal/std Max 5.19766 +trainer/policy/normal/std Min 0.288347 +trainer/policy/normal/log_std Mean 0.988621 +trainer/policy/normal/log_std Std 0.313076 +trainer/policy/normal/log_std Max 1.64821 +trainer/policy/normal/log_std Min -1.24359 +eval/num steps total 577133 +eval/num paths total 579 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.112965 +eval/Actions Std 0.9063 +eval/Actions Max 0.999981 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31999 +time/logging (s) 0.00423789 +time/sampling batch (s) 0.529711 +time/saving (s) 0.00405509 +time/training (s) 7.03672 +time/epoch (s) 9.89472 +time/total (s) 6009.02 +Epoch -422 +---------------------------------- --------------- +2022-05-10 14:51:03.098659 PDT | [1] Epoch -421 finished +---------------------------------- --------------- +epoch -421 +replay_buffer/size 999033 +trainer/num train calls 580000 +trainer/Policy Loss -18.2522 +trainer/Log Pis Mean 24.2339 +trainer/Log Pis Std 13.9982 +trainer/Log Pis Max 69.3365 +trainer/Log Pis Min -10.3402 +trainer/policy/mean Mean -0.0256362 +trainer/policy/mean Std 0.904601 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85918 +trainer/policy/normal/std Std 0.671272 +trainer/policy/normal/std Max 6.6346 +trainer/policy/normal/std Min 0.337738 +trainer/policy/normal/log_std Mean 1.01223 +trainer/policy/normal/log_std Std 0.309116 +trainer/policy/normal/log_std Max 1.8923 +trainer/policy/normal/log_std Min -1.08548 +eval/num steps total 578133 +eval/num paths total 580 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0725602 +eval/Actions Std 0.91244 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6227 +time/logging (s) 0.00414019 +time/sampling batch (s) 0.533883 +time/saving (s) 0.00394652 +time/training (s) 6.8466 +time/epoch (s) 10.0113 +time/total (s) 6019.03 +Epoch -421 +---------------------------------- --------------- +2022-05-10 14:51:13.313757 PDT | [1] Epoch -420 finished +---------------------------------- --------------- +epoch -420 +replay_buffer/size 999033 +trainer/num train calls 581000 +trainer/Policy Loss -20.5351 +trainer/Log Pis Mean 24.5491 +trainer/Log Pis Std 12.5461 +trainer/Log Pis Max 74.1732 +trainer/Log Pis Min -6.0627 +trainer/policy/mean Mean -0.0495768 +trainer/policy/mean Std 0.907518 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83421 +trainer/policy/normal/std Std 0.628668 +trainer/policy/normal/std Max 5.63236 +trainer/policy/normal/std Min 0.37718 +trainer/policy/normal/log_std Mean 1.00811 +trainer/policy/normal/log_std Std 0.288428 +trainer/policy/normal/log_std Max 1.72853 +trainer/policy/normal/log_std Min -0.975032 +eval/num steps total 579133 +eval/num paths total 581 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.354994 +eval/Actions Std 0.833097 +eval/Actions Max 0.999994 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5378 +time/logging (s) 0.00387905 +time/sampling batch (s) 0.526912 +time/saving (s) 0.00371653 +time/training (s) 7.11906 +time/epoch (s) 10.1914 +time/total (s) 6029.23 +Epoch -420 +---------------------------------- --------------- +2022-05-10 14:51:23.444403 PDT | [1] Epoch -419 finished +---------------------------------- --------------- +epoch -419 +replay_buffer/size 999033 +trainer/num train calls 582000 +trainer/Policy Loss -18.4887 +trainer/Log Pis Mean 23.8116 +trainer/Log Pis Std 14.3191 +trainer/Log Pis Max 65.9997 +trainer/Log Pis Min -11.0803 +trainer/policy/mean Mean -0.0305896 +trainer/policy/mean Std 0.904461 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87087 +trainer/policy/normal/std Std 0.665477 +trainer/policy/normal/std Max 6.67128 +trainer/policy/normal/std Min 0.306391 +trainer/policy/normal/log_std Mean 1.01615 +trainer/policy/normal/log_std Std 0.312803 +trainer/policy/normal/log_std Max 1.89781 +trainer/policy/normal/log_std Min -1.18289 +eval/num steps total 580133 +eval/num paths total 582 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.149273 +eval/Actions Std 0.89566 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77281 +time/logging (s) 0.00379031 +time/sampling batch (s) 0.284396 +time/saving (s) 0.00348335 +time/training (s) 7.04262 +time/epoch (s) 10.1071 +time/total (s) 6039.34 +Epoch -419 +---------------------------------- --------------- +2022-05-10 14:51:33.465563 PDT | [1] Epoch -418 finished +---------------------------------- --------------- +epoch -418 +replay_buffer/size 999033 +trainer/num train calls 583000 +trainer/Policy Loss -19.6946 +trainer/Log Pis Mean 24.3887 +trainer/Log Pis Std 13.4964 +trainer/Log Pis Max 68.3774 +trainer/Log Pis Min -9.47844 +trainer/policy/mean Mean -0.0404084 +trainer/policy/mean Std 0.905834 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79571 +trainer/policy/normal/std Std 0.628049 +trainer/policy/normal/std Max 5.88867 +trainer/policy/normal/std Min 0.341863 +trainer/policy/normal/log_std Mean 0.994226 +trainer/policy/normal/log_std Std 0.287721 +trainer/policy/normal/log_std Max 1.77303 +trainer/policy/normal/log_std Min -1.07334 +eval/num steps total 581133 +eval/num paths total 583 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.343089 +eval/Actions Std 0.891612 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81263 +time/logging (s) 0.00458255 +time/sampling batch (s) 0.536445 +time/saving (s) 0.00452361 +time/training (s) 6.64009 +time/epoch (s) 9.99827 +time/total (s) 6049.34 +Epoch -418 +---------------------------------- --------------- +2022-05-10 14:51:43.471954 PDT | [1] Epoch -417 finished +---------------------------------- --------------- +epoch -417 +replay_buffer/size 999033 +trainer/num train calls 584000 +trainer/Policy Loss -20.2791 +trainer/Log Pis Mean 25.0657 +trainer/Log Pis Std 13.1999 +trainer/Log Pis Max 73.7437 +trainer/Log Pis Min -9.61913 +trainer/policy/mean Mean -0.0288722 +trainer/policy/mean Std 0.908657 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81671 +trainer/policy/normal/std Std 0.667913 +trainer/policy/normal/std Max 6.32545 +trainer/policy/normal/std Min 0.286564 +trainer/policy/normal/log_std Mean 0.99571 +trainer/policy/normal/log_std Std 0.317936 +trainer/policy/normal/log_std Max 1.84458 +trainer/policy/normal/log_std Min -1.24979 +eval/num steps total 582133 +eval/num paths total 584 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0444212 +eval/Actions Std 0.91283 +eval/Actions Max 0.999994 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7507 +time/logging (s) 0.0036858 +time/sampling batch (s) 0.278271 +time/saving (s) 0.00342096 +time/training (s) 6.94463 +time/epoch (s) 9.98071 +time/total (s) 6059.32 +Epoch -417 +---------------------------------- --------------- +2022-05-10 14:51:53.653099 PDT | [1] Epoch -416 finished +---------------------------------- --------------- +epoch -416 +replay_buffer/size 999033 +trainer/num train calls 585000 +trainer/Policy Loss -20.7453 +trainer/Log Pis Mean 25.9875 +trainer/Log Pis Std 13.2429 +trainer/Log Pis Max 65.8181 +trainer/Log Pis Min -7.18638 +trainer/policy/mean Mean -0.0335742 +trainer/policy/mean Std 0.910876 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.79946 +trainer/policy/normal/std Std 0.652012 +trainer/policy/normal/std Max 7.35986 +trainer/policy/normal/std Min 0.281313 +trainer/policy/normal/log_std Mean 0.990369 +trainer/policy/normal/log_std Std 0.315473 +trainer/policy/normal/log_std Max 1.99604 +trainer/policy/normal/log_std Min -1.26829 +eval/num steps total 583133 +eval/num paths total 585 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.335541 +eval/Actions Std 0.71469 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75986 +time/logging (s) 0.00415503 +time/sampling batch (s) 0.275205 +time/saving (s) 0.004127 +time/training (s) 7.11538 +time/epoch (s) 10.1587 +time/total (s) 6069.48 +Epoch -416 +---------------------------------- --------------- +2022-05-10 14:52:03.070473 PDT | [1] Epoch -415 finished +---------------------------------- --------------- +epoch -415 +replay_buffer/size 999033 +trainer/num train calls 586000 +trainer/Policy Loss -19.7909 +trainer/Log Pis Mean 25.5341 +trainer/Log Pis Std 13.1047 +trainer/Log Pis Max 71.7847 +trainer/Log Pis Min -8.24383 +trainer/policy/mean Mean -0.0245029 +trainer/policy/mean Std 0.909149 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80309 +trainer/policy/normal/std Std 0.66102 +trainer/policy/normal/std Max 6.52636 +trainer/policy/normal/std Min 0.256507 +trainer/policy/normal/log_std Mean 0.989526 +trainer/policy/normal/log_std Std 0.3273 +trainer/policy/normal/log_std Max 1.87585 +trainer/policy/normal/log_std Min -1.3606 +eval/num steps total 584133 +eval/num paths total 586 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111381 +eval/Actions Std 0.911652 +eval/Actions Max 0.999992 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49335 +time/logging (s) 0.00367605 +time/sampling batch (s) 0.775686 +time/saving (s) 0.00338727 +time/training (s) 6.11742 +time/epoch (s) 9.39353 +time/total (s) 6078.88 +Epoch -415 +---------------------------------- --------------- +2022-05-10 14:52:12.683001 PDT | [1] Epoch -414 finished +---------------------------------- --------------- +epoch -414 +replay_buffer/size 999033 +trainer/num train calls 587000 +trainer/Policy Loss -19.276 +trainer/Log Pis Mean 24.2544 +trainer/Log Pis Std 13.3852 +trainer/Log Pis Max 65.3597 +trainer/Log Pis Min -10.9293 +trainer/policy/mean Mean -0.0167248 +trainer/policy/mean Std 0.907212 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77993 +trainer/policy/normal/std Std 0.657401 +trainer/policy/normal/std Max 5.85435 +trainer/policy/normal/std Min 0.296832 +trainer/policy/normal/log_std Mean 0.981312 +trainer/policy/normal/log_std Std 0.325333 +trainer/policy/normal/log_std Max 1.76718 +trainer/policy/normal/log_std Min -1.21459 +eval/num steps total 585133 +eval/num paths total 587 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.212213 +eval/Actions Std 0.903205 +eval/Actions Max 0.999999 +eval/Actions Min -0.999971 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46101 +time/logging (s) 0.00370563 +time/sampling batch (s) 0.275184 +time/saving (s) 0.0034282 +time/training (s) 6.84643 +time/epoch (s) 9.58976 +time/total (s) 6088.47 +Epoch -414 +---------------------------------- --------------- +2022-05-10 14:52:22.882557 PDT | [1] Epoch -413 finished +---------------------------------- --------------- +epoch -413 +replay_buffer/size 999033 +trainer/num train calls 588000 +trainer/Policy Loss -20.4712 +trainer/Log Pis Mean 24.9314 +trainer/Log Pis Std 13.2733 +trainer/Log Pis Max 66.1372 +trainer/Log Pis Min -11.5371 +trainer/policy/mean Mean -0.0297317 +trainer/policy/mean Std 0.905745 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81294 +trainer/policy/normal/std Std 0.6517 +trainer/policy/normal/std Max 5.51812 +trainer/policy/normal/std Min 0.209398 +trainer/policy/normal/log_std Mean 0.994311 +trainer/policy/normal/log_std Std 0.32311 +trainer/policy/normal/log_std Max 1.70804 +trainer/policy/normal/log_std Min -1.56352 +eval/num steps total 586133 +eval/num paths total 588 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0989002 +eval/Actions Std 0.911189 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6749 +time/logging (s) 0.00405876 +time/sampling batch (s) 0.276156 +time/saving (s) 0.00378289 +time/training (s) 7.21804 +time/epoch (s) 10.1769 +time/total (s) 6098.65 +Epoch -413 +---------------------------------- --------------- +2022-05-10 14:52:34.400720 PDT | [1] Epoch -412 finished +---------------------------------- --------------- +epoch -412 +replay_buffer/size 999033 +trainer/num train calls 589000 +trainer/Policy Loss -19.1328 +trainer/Log Pis Mean 24.296 +trainer/Log Pis Std 12.7431 +trainer/Log Pis Max 69.0866 +trainer/Log Pis Min -5.77798 +trainer/policy/mean Mean -0.0450285 +trainer/policy/mean Std 0.905264 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77503 +trainer/policy/normal/std Std 0.641783 +trainer/policy/normal/std Max 6.28727 +trainer/policy/normal/std Min 0.294937 +trainer/policy/normal/log_std Mean 0.981271 +trainer/policy/normal/log_std Std 0.319138 +trainer/policy/normal/log_std Max 1.83853 +trainer/policy/normal/log_std Min -1.22099 +eval/num steps total 587133 +eval/num paths total 589 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.287642 +eval/Actions Std 0.721668 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66121 +time/logging (s) 0.00382869 +time/sampling batch (s) 0.525371 +time/saving (s) 0.00368096 +time/training (s) 8.30045 +time/epoch (s) 11.4945 +time/total (s) 6110.15 +Epoch -412 +---------------------------------- --------------- +2022-05-10 14:52:44.632755 PDT | [1] Epoch -411 finished +---------------------------------- --------------- +epoch -411 +replay_buffer/size 999033 +trainer/num train calls 590000 +trainer/Policy Loss -20.2752 +trainer/Log Pis Mean 23.797 +trainer/Log Pis Std 13.0311 +trainer/Log Pis Max 67.0182 +trainer/Log Pis Min -6.846 +trainer/policy/mean Mean -0.0426357 +trainer/policy/mean Std 0.908039 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79171 +trainer/policy/normal/std Std 0.640357 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.326195 +trainer/policy/normal/log_std Mean 0.989183 +trainer/policy/normal/log_std Std 0.308285 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.12026 +eval/num steps total 588133 +eval/num paths total 590 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0808947 +eval/Actions Std 0.910196 +eval/Actions Max 0.999986 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55731 +time/logging (s) 0.00366631 +time/sampling batch (s) 0.776223 +time/saving (s) 0.00344555 +time/training (s) 6.86823 +time/epoch (s) 10.2089 +time/total (s) 6120.36 +Epoch -411 +---------------------------------- --------------- +2022-05-10 14:52:54.918156 PDT | [1] Epoch -410 finished +---------------------------------- --------------- +epoch -410 +replay_buffer/size 999033 +trainer/num train calls 591000 +trainer/Policy Loss -21.2794 +trainer/Log Pis Mean 24.4241 +trainer/Log Pis Std 13.3907 +trainer/Log Pis Max 79.5479 +trainer/Log Pis Min -5.97583 +trainer/policy/mean Mean -0.0365304 +trainer/policy/mean Std 0.907373 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80007 +trainer/policy/normal/std Std 0.633932 +trainer/policy/normal/std Max 6.44613 +trainer/policy/normal/std Min 0.361676 +trainer/policy/normal/log_std Mean 0.993782 +trainer/policy/normal/log_std Std 0.30002 +trainer/policy/normal/log_std Max 1.86348 +trainer/policy/normal/log_std Min -1.01701 +eval/num steps total 589133 +eval/num paths total 591 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0170316 +eval/Actions Std 0.926477 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35816 +time/logging (s) 0.00369269 +time/sampling batch (s) 1.03086 +time/saving (s) 0.00346517 +time/training (s) 6.86615 +time/epoch (s) 10.2623 +time/total (s) 6130.63 +Epoch -410 +---------------------------------- --------------- +2022-05-10 14:53:06.580108 PDT | [1] Epoch -409 finished +---------------------------------- --------------- +epoch -409 +replay_buffer/size 999033 +trainer/num train calls 592000 +trainer/Policy Loss -19.0584 +trainer/Log Pis Mean 25.1095 +trainer/Log Pis Std 12.8824 +trainer/Log Pis Max 66.2836 +trainer/Log Pis Min -4.12378 +trainer/policy/mean Mean -0.0551398 +trainer/policy/mean Std 0.903596 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.81677 +trainer/policy/normal/std Std 0.655141 +trainer/policy/normal/std Max 5.75755 +trainer/policy/normal/std Min 0.283425 +trainer/policy/normal/log_std Mean 0.996615 +trainer/policy/normal/log_std Std 0.315318 +trainer/policy/normal/log_std Max 1.75051 +trainer/policy/normal/log_std Min -1.26081 +eval/num steps total 590133 +eval/num paths total 592 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.205254 +eval/Actions Std 0.885318 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6893 +time/logging (s) 0.00415238 +time/sampling batch (s) 0.530567 +time/saving (s) 0.00401488 +time/training (s) 8.41118 +time/epoch (s) 11.6392 +time/total (s) 6142.27 +Epoch -409 +---------------------------------- --------------- +2022-05-10 14:53:16.932534 PDT | [1] Epoch -408 finished +---------------------------------- --------------- +epoch -408 +replay_buffer/size 999033 +trainer/num train calls 593000 +trainer/Policy Loss -19.4133 +trainer/Log Pis Mean 24.3762 +trainer/Log Pis Std 12.6696 +trainer/Log Pis Max 65.6659 +trainer/Log Pis Min -5.74547 +trainer/policy/mean Mean -0.019878 +trainer/policy/mean Std 0.909214 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81234 +trainer/policy/normal/std Std 0.644037 +trainer/policy/normal/std Max 6.36002 +trainer/policy/normal/std Min 0.311854 +trainer/policy/normal/log_std Mean 0.995331 +trainer/policy/normal/log_std Std 0.316736 +trainer/policy/normal/log_std Max 1.85003 +trainer/policy/normal/log_std Min -1.16522 +eval/num steps total 591133 +eval/num paths total 593 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110672 +eval/Actions Std 0.863208 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54167 +time/logging (s) 0.00374201 +time/sampling batch (s) 0.52902 +time/saving (s) 0.00368617 +time/training (s) 7.25039 +time/epoch (s) 10.3285 +time/total (s) 6152.6 +Epoch -408 +---------------------------------- --------------- +2022-05-10 14:53:27.607720 PDT | [1] Epoch -407 finished +---------------------------------- --------------- +epoch -407 +replay_buffer/size 999033 +trainer/num train calls 594000 +trainer/Policy Loss -19.3421 +trainer/Log Pis Mean 24.1702 +trainer/Log Pis Std 13.0306 +trainer/Log Pis Max 81.5664 +trainer/Log Pis Min -9.43236 +trainer/policy/mean Mean -0.0424242 +trainer/policy/mean Std 0.904364 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79275 +trainer/policy/normal/std Std 0.666624 +trainer/policy/normal/std Max 5.90711 +trainer/policy/normal/std Min 0.246459 +trainer/policy/normal/log_std Mean 0.985316 +trainer/policy/normal/log_std Std 0.329134 +trainer/policy/normal/log_std Max 1.77616 +trainer/policy/normal/log_std Min -1.40056 +eval/num steps total 592133 +eval/num paths total 594 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.359789 +eval/Actions Std 0.767267 +eval/Actions Max 0.99999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59532 +time/logging (s) 0.00369806 +time/sampling batch (s) 0.284176 +time/saving (s) 0.00347466 +time/training (s) 7.76511 +time/epoch (s) 10.6518 +time/total (s) 6163.26 +Epoch -407 +---------------------------------- --------------- +2022-05-10 14:53:38.042939 PDT | [1] Epoch -406 finished +---------------------------------- --------------- +epoch -406 +replay_buffer/size 999033 +trainer/num train calls 595000 +trainer/Policy Loss -19.6816 +trainer/Log Pis Mean 24.1782 +trainer/Log Pis Std 12.8802 +trainer/Log Pis Max 71.3585 +trainer/Log Pis Min -5.10874 +trainer/policy/mean Mean -0.0343765 +trainer/policy/mean Std 0.905685 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76376 +trainer/policy/normal/std Std 0.628817 +trainer/policy/normal/std Max 5.30147 +trainer/policy/normal/std Min 0.254119 +trainer/policy/normal/log_std Mean 0.978976 +trainer/policy/normal/log_std Std 0.30971 +trainer/policy/normal/log_std Max 1.66798 +trainer/policy/normal/log_std Min -1.36995 +eval/num steps total 593133 +eval/num paths total 595 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.29548 +eval/Actions Std 0.877362 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4892 +time/logging (s) 0.00372643 +time/sampling batch (s) 0.27951 +time/saving (s) 0.00354962 +time/training (s) 7.63627 +time/epoch (s) 10.4123 +time/total (s) 6173.67 +Epoch -406 +---------------------------------- --------------- +2022-05-10 14:53:48.740572 PDT | [1] Epoch -405 finished +---------------------------------- --------------- +epoch -405 +replay_buffer/size 999033 +trainer/num train calls 596000 +trainer/Policy Loss -19.6386 +trainer/Log Pis Mean 24.2979 +trainer/Log Pis Std 13.4479 +trainer/Log Pis Max 74.3445 +trainer/Log Pis Min -9.37609 +trainer/policy/mean Mean -0.0199264 +trainer/policy/mean Std 0.904857 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.72259 +trainer/policy/normal/std Std 0.640371 +trainer/policy/normal/std Max 5.94683 +trainer/policy/normal/std Min 0.225845 +trainer/policy/normal/log_std Mean 0.960839 +trainer/policy/normal/log_std Std 0.323935 +trainer/policy/normal/log_std Max 1.78286 +trainer/policy/normal/log_std Min -1.48791 +eval/num steps total 594133 +eval/num paths total 596 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0375539 +eval/Actions Std 0.901784 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56317 +time/logging (s) 0.00522061 +time/sampling batch (s) 0.347203 +time/saving (s) 0.00495801 +time/training (s) 7.75397 +time/epoch (s) 10.6745 +time/total (s) 6184.35 +Epoch -405 +---------------------------------- --------------- +2022-05-10 14:53:59.364915 PDT | [1] Epoch -404 finished +---------------------------------- --------------- +epoch -404 +replay_buffer/size 999033 +trainer/num train calls 597000 +trainer/Policy Loss -19.4756 +trainer/Log Pis Mean 24.065 +trainer/Log Pis Std 13.5125 +trainer/Log Pis Max 72.7756 +trainer/Log Pis Min -8.16148 +trainer/policy/mean Mean -0.0339302 +trainer/policy/mean Std 0.906498 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85442 +trainer/policy/normal/std Std 0.661344 +trainer/policy/normal/std Max 6.5718 +trainer/policy/normal/std Min 0.268421 +trainer/policy/normal/log_std Mean 1.01093 +trainer/policy/normal/log_std Std 0.309711 +trainer/policy/normal/log_std Max 1.88279 +trainer/policy/normal/log_std Min -1.3152 +eval/num steps total 595133 +eval/num paths total 597 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.109914 +eval/Actions Std 0.831051 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74006 +time/logging (s) 0.00376784 +time/sampling batch (s) 0.561536 +time/saving (s) 0.00349294 +time/training (s) 7.28863 +time/epoch (s) 10.5975 +time/total (s) 6194.95 +Epoch -404 +---------------------------------- --------------- +2022-05-10 14:54:09.537174 PDT | [1] Epoch -403 finished +---------------------------------- --------------- +epoch -403 +replay_buffer/size 999033 +trainer/num train calls 598000 +trainer/Policy Loss -19.3482 +trainer/Log Pis Mean 24.0822 +trainer/Log Pis Std 13.0977 +trainer/Log Pis Max 68.1363 +trainer/Log Pis Min -8.071 +trainer/policy/mean Mean -0.0337474 +trainer/policy/mean Std 0.903576 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78094 +trainer/policy/normal/std Std 0.649914 +trainer/policy/normal/std Max 6.27946 +trainer/policy/normal/std Min 0.246 +trainer/policy/normal/log_std Mean 0.984802 +trainer/policy/normal/log_std Std 0.308599 +trainer/policy/normal/log_std Max 1.83728 +trainer/policy/normal/log_std Min -1.40242 +eval/num steps total 596133 +eval/num paths total 598 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.225607 +eval/Actions Std 0.94207 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65009 +time/logging (s) 0.00376291 +time/sampling batch (s) 0.275977 +time/saving (s) 0.00349652 +time/training (s) 7.21604 +time/epoch (s) 10.1494 +time/total (s) 6205.1 +Epoch -403 +---------------------------------- --------------- +2022-05-10 14:54:19.786307 PDT | [1] Epoch -402 finished +---------------------------------- ---------------- +epoch -402 +replay_buffer/size 999033 +trainer/num train calls 599000 +trainer/Policy Loss -18.254 +trainer/Log Pis Mean 23.9808 +trainer/Log Pis Std 12.9507 +trainer/Log Pis Max 74.0766 +trainer/Log Pis Min -4.64354 +trainer/policy/mean Mean -0.0247262 +trainer/policy/mean Std 0.905906 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.73437 +trainer/policy/normal/std Std 0.657799 +trainer/policy/normal/std Max 5.92406 +trainer/policy/normal/std Min 0.28789 +trainer/policy/normal/log_std Mean 0.962871 +trainer/policy/normal/log_std Std 0.335074 +trainer/policy/normal/log_std Max 1.77902 +trainer/policy/normal/log_std Min -1.24518 +eval/num steps total 597133 +eval/num paths total 599 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000232683 +eval/Actions Std 0.907748 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63389 +time/logging (s) 0.00403549 +time/sampling batch (s) 0.526478 +time/saving (s) 0.00384627 +time/training (s) 7.05803 +time/epoch (s) 10.2263 +time/total (s) 6215.33 +Epoch -402 +---------------------------------- ---------------- +2022-05-10 14:54:30.041267 PDT | [1] Epoch -401 finished +---------------------------------- ---------------- +epoch -401 +replay_buffer/size 999033 +trainer/num train calls 600000 +trainer/Policy Loss -19.7852 +trainer/Log Pis Mean 24.6898 +trainer/Log Pis Std 12.8859 +trainer/Log Pis Max 69.7116 +trainer/Log Pis Min -3.47443 +trainer/policy/mean Mean -0.0233509 +trainer/policy/mean Std 0.907992 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82934 +trainer/policy/normal/std Std 0.629466 +trainer/policy/normal/std Max 6.07007 +trainer/policy/normal/std Min 0.332148 +trainer/policy/normal/log_std Mean 1.00527 +trainer/policy/normal/log_std Std 0.295507 +trainer/policy/normal/log_std Max 1.80337 +trainer/policy/normal/log_std Min -1.10217 +eval/num steps total 598133 +eval/num paths total 600 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000488248 +eval/Actions Std 0.909879 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54006 +time/logging (s) 0.00374701 +time/sampling batch (s) 0.528288 +time/saving (s) 0.00652672 +time/training (s) 7.15266 +time/epoch (s) 10.2313 +time/total (s) 6225.57 +Epoch -401 +---------------------------------- ---------------- +2022-05-10 14:54:39.831277 PDT | [1] Epoch -400 finished +---------------------------------- --------------- +epoch -400 +replay_buffer/size 999033 +trainer/num train calls 601000 +trainer/Policy Loss -20.3712 +trainer/Log Pis Mean 24.316 +trainer/Log Pis Std 13.2592 +trainer/Log Pis Max 60.0002 +trainer/Log Pis Min -5.40321 +trainer/policy/mean Mean -0.0412315 +trainer/policy/mean Std 0.903343 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86048 +trainer/policy/normal/std Std 0.646079 +trainer/policy/normal/std Max 5.46401 +trainer/policy/normal/std Min 0.273062 +trainer/policy/normal/log_std Mean 1.01463 +trainer/policy/normal/log_std Std 0.303759 +trainer/policy/normal/log_std Max 1.69818 +trainer/policy/normal/log_std Min -1.29806 +eval/num steps total 599133 +eval/num paths total 601 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.558982 +eval/Actions Std 0.705169 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75895 +time/logging (s) 0.00382352 +time/sampling batch (s) 0.281297 +time/saving (s) 0.00352403 +time/training (s) 6.71937 +time/epoch (s) 9.76696 +time/total (s) 6235.34 +Epoch -400 +---------------------------------- --------------- +2022-05-10 14:54:49.361745 PDT | [1] Epoch -399 finished +---------------------------------- --------------- +epoch -399 +replay_buffer/size 999033 +trainer/num train calls 602000 +trainer/Policy Loss -18.9689 +trainer/Log Pis Mean 24.0065 +trainer/Log Pis Std 12.6896 +trainer/Log Pis Max 62.1432 +trainer/Log Pis Min -8.44181 +trainer/policy/mean Mean -0.0266537 +trainer/policy/mean Std 0.905546 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8173 +trainer/policy/normal/std Std 0.652831 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.339519 +trainer/policy/normal/log_std Mean 0.996946 +trainer/policy/normal/log_std Std 0.315392 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.08022 +eval/num steps total 600133 +eval/num paths total 602 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133461 +eval/Actions Std 0.904492 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5927 +time/logging (s) 0.00456197 +time/sampling batch (s) 0.281931 +time/saving (s) 0.0041632 +time/training (s) 6.62452 +time/epoch (s) 9.50788 +time/total (s) 6244.85 +Epoch -399 +---------------------------------- --------------- +2022-05-10 14:54:59.718250 PDT | [1] Epoch -398 finished +---------------------------------- --------------- +epoch -398 +replay_buffer/size 999033 +trainer/num train calls 603000 +trainer/Policy Loss -20.3132 +trainer/Log Pis Mean 24.7314 +trainer/Log Pis Std 13.0392 +trainer/Log Pis Max 72.278 +trainer/Log Pis Min -6.03006 +trainer/policy/mean Mean -0.0460793 +trainer/policy/mean Std 0.899021 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83857 +trainer/policy/normal/std Std 0.693727 +trainer/policy/normal/std Max 5.90423 +trainer/policy/normal/std Min 0.203544 +trainer/policy/normal/log_std Mean 0.996537 +trainer/policy/normal/log_std Std 0.354818 +trainer/policy/normal/log_std Max 1.77567 +trainer/policy/normal/log_std Min -1.59187 +eval/num steps total 601133 +eval/num paths total 603 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115164 +eval/Actions Std 0.895705 +eval/Actions Max 0.999994 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79998 +time/logging (s) 0.00407054 +time/sampling batch (s) 0.281974 +time/saving (s) 0.00395436 +time/training (s) 7.24189 +time/epoch (s) 10.3319 +time/total (s) 6255.18 +Epoch -398 +---------------------------------- --------------- +2022-05-10 14:55:09.786067 PDT | [1] Epoch -397 finished +---------------------------------- --------------- +epoch -397 +replay_buffer/size 999033 +trainer/num train calls 604000 +trainer/Policy Loss -19.0805 +trainer/Log Pis Mean 24.8026 +trainer/Log Pis Std 13.5603 +trainer/Log Pis Max 75.9636 +trainer/Log Pis Min -2.67018 +trainer/policy/mean Mean -0.0510161 +trainer/policy/mean Std 0.903155 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83391 +trainer/policy/normal/std Std 0.667877 +trainer/policy/normal/std Max 5.66541 +trainer/policy/normal/std Min 0.293948 +trainer/policy/normal/log_std Mean 1.00159 +trainer/policy/normal/log_std Std 0.319435 +trainer/policy/normal/log_std Max 1.73438 +trainer/policy/normal/log_std Min -1.22435 +eval/num steps total 602133 +eval/num paths total 604 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125606 +eval/Actions Std 0.914504 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67357 +time/logging (s) 0.00428295 +time/sampling batch (s) 0.548555 +time/saving (s) 0.00413408 +time/training (s) 6.81306 +time/epoch (s) 10.0436 +time/total (s) 6265.23 +Epoch -397 +---------------------------------- --------------- +2022-05-10 14:55:19.911077 PDT | [1] Epoch -396 finished +---------------------------------- --------------- +epoch -396 +replay_buffer/size 999033 +trainer/num train calls 605000 +trainer/Policy Loss -18.9856 +trainer/Log Pis Mean 25.6753 +trainer/Log Pis Std 13.6579 +trainer/Log Pis Max 74.3849 +trainer/Log Pis Min -6.50952 +trainer/policy/mean Mean -0.0166685 +trainer/policy/mean Std 0.909329 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78882 +trainer/policy/normal/std Std 0.649572 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.2393 +trainer/policy/normal/log_std Mean 0.985463 +trainer/policy/normal/log_std Std 0.323052 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.43004 +eval/num steps total 603133 +eval/num paths total 605 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121741 +eval/Actions Std 0.917715 +eval/Actions Max 0.999996 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62475 +time/logging (s) 0.00427406 +time/sampling batch (s) 0.2782 +time/saving (s) 0.00414642 +time/training (s) 7.18979 +time/epoch (s) 10.1012 +time/total (s) 6275.34 +Epoch -396 +---------------------------------- --------------- +2022-05-10 14:55:29.517052 PDT | [1] Epoch -395 finished +---------------------------------- --------------- +epoch -395 +replay_buffer/size 999033 +trainer/num train calls 606000 +trainer/Policy Loss -19.2518 +trainer/Log Pis Mean 24.5128 +trainer/Log Pis Std 13.6265 +trainer/Log Pis Max 75.2236 +trainer/Log Pis Min -6.90016 +trainer/policy/mean Mean -0.0355707 +trainer/policy/mean Std 0.909015 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80652 +trainer/policy/normal/std Std 0.640412 +trainer/policy/normal/std Max 6.20467 +trainer/policy/normal/std Min 0.275107 +trainer/policy/normal/log_std Mean 0.994008 +trainer/policy/normal/log_std Std 0.312123 +trainer/policy/normal/log_std Max 1.8253 +trainer/policy/normal/log_std Min -1.29059 +eval/num steps total 604133 +eval/num paths total 606 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0745993 +eval/Actions Std 0.866426 +eval/Actions Max 0.999999 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62357 +time/logging (s) 0.00424163 +time/sampling batch (s) 0.281193 +time/saving (s) 0.00405491 +time/training (s) 6.66896 +time/epoch (s) 9.58202 +time/total (s) 6284.92 +Epoch -395 +---------------------------------- --------------- +2022-05-10 14:55:40.279634 PDT | [1] Epoch -394 finished +---------------------------------- --------------- +epoch -394 +replay_buffer/size 999033 +trainer/num train calls 607000 +trainer/Policy Loss -19.1237 +trainer/Log Pis Mean 23.7024 +trainer/Log Pis Std 12.9986 +trainer/Log Pis Max 65.7596 +trainer/Log Pis Min -9.56218 +trainer/policy/mean Mean -0.0462829 +trainer/policy/mean Std 0.906705 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77725 +trainer/policy/normal/std Std 0.64264 +trainer/policy/normal/std Max 5.92909 +trainer/policy/normal/std Min 0.262134 +trainer/policy/normal/log_std Mean 0.982428 +trainer/policy/normal/log_std Std 0.316952 +trainer/policy/normal/log_std Max 1.77987 +trainer/policy/normal/log_std Min -1.3389 +eval/num steps total 605133 +eval/num paths total 607 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.147891 +eval/Actions Std 0.878895 +eval/Actions Max 0.999985 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69024 +time/logging (s) 0.00374183 +time/sampling batch (s) 0.530369 +time/saving (s) 0.00352921 +time/training (s) 7.51039 +time/epoch (s) 10.7383 +time/total (s) 6295.66 +Epoch -394 +---------------------------------- --------------- +2022-05-10 14:55:50.084212 PDT | [1] Epoch -393 finished +---------------------------------- --------------- +epoch -393 +replay_buffer/size 999033 +trainer/num train calls 608000 +trainer/Policy Loss -20.3146 +trainer/Log Pis Mean 24.1653 +trainer/Log Pis Std 13.2139 +trainer/Log Pis Max 68.5071 +trainer/Log Pis Min -10.3696 +trainer/policy/mean Mean -0.0266253 +trainer/policy/mean Std 0.903582 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74639 +trainer/policy/normal/std Std 0.655004 +trainer/policy/normal/std Max 5.11765 +trainer/policy/normal/std Min 0.249471 +trainer/policy/normal/log_std Mean 0.967781 +trainer/policy/normal/log_std Std 0.3323 +trainer/policy/normal/log_std Max 1.6327 +trainer/policy/normal/log_std Min -1.38841 +eval/num steps total 606133 +eval/num paths total 608 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0444601 +eval/Actions Std 0.904081 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.29055 +time/logging (s) 0.00427835 +time/sampling batch (s) 0.278517 +time/saving (s) 0.00397528 +time/training (s) 7.20442 +time/epoch (s) 9.78175 +time/total (s) 6305.45 +Epoch -393 +---------------------------------- --------------- +2022-05-10 14:56:00.755976 PDT | [1] Epoch -392 finished +---------------------------------- --------------- +epoch -392 +replay_buffer/size 999033 +trainer/num train calls 609000 +trainer/Policy Loss -20.2165 +trainer/Log Pis Mean 23.793 +trainer/Log Pis Std 13.6384 +trainer/Log Pis Max 67.9743 +trainer/Log Pis Min -10.8611 +trainer/policy/mean Mean -0.0488437 +trainer/policy/mean Std 0.905418 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85454 +trainer/policy/normal/std Std 0.664252 +trainer/policy/normal/std Max 6.39082 +trainer/policy/normal/std Min 0.298488 +trainer/policy/normal/log_std Mean 1.01073 +trainer/policy/normal/log_std Std 0.310311 +trainer/policy/normal/log_std Max 1.85486 +trainer/policy/normal/log_std Min -1.20903 +eval/num steps total 607133 +eval/num paths total 609 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.226008 +eval/Actions Std 0.945372 +eval/Actions Max 0.999996 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78491 +time/logging (s) 0.00392125 +time/sampling batch (s) 0.525922 +time/saving (s) 0.00352653 +time/training (s) 7.32988 +time/epoch (s) 10.6482 +time/total (s) 6316.1 +Epoch -392 +---------------------------------- --------------- +2022-05-10 14:56:11.867640 PDT | [1] Epoch -391 finished +---------------------------------- --------------- +epoch -391 +replay_buffer/size 999033 +trainer/num train calls 610000 +trainer/Policy Loss -18.2712 +trainer/Log Pis Mean 23.3188 +trainer/Log Pis Std 13.4691 +trainer/Log Pis Max 75.4924 +trainer/Log Pis Min -6.39826 +trainer/policy/mean Mean -0.0427367 +trainer/policy/mean Std 0.902852 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77675 +trainer/policy/normal/std Std 0.669692 +trainer/policy/normal/std Max 6.36317 +trainer/policy/normal/std Min 0.22065 +trainer/policy/normal/log_std Mean 0.978412 +trainer/policy/normal/log_std Std 0.333163 +trainer/policy/normal/log_std Max 1.85053 +trainer/policy/normal/log_std Min -1.51118 +eval/num steps total 608133 +eval/num paths total 610 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0454116 +eval/Actions Std 0.895391 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70643 +time/logging (s) 0.0038941 +time/sampling batch (s) 0.776194 +time/saving (s) 0.00344399 +time/training (s) 7.59874 +time/epoch (s) 11.0887 +time/total (s) 6327.19 +Epoch -391 +---------------------------------- --------------- +2022-05-10 14:56:22.759671 PDT | [1] Epoch -390 finished +---------------------------------- --------------- +epoch -390 +replay_buffer/size 999033 +trainer/num train calls 611000 +trainer/Policy Loss -19.5397 +trainer/Log Pis Mean 24.8069 +trainer/Log Pis Std 13.5652 +trainer/Log Pis Max 66.9902 +trainer/Log Pis Min -7.6029 +trainer/policy/mean Mean -0.0338591 +trainer/policy/mean Std 0.911743 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80662 +trainer/policy/normal/std Std 0.627464 +trainer/policy/normal/std Max 5.54876 +trainer/policy/normal/std Min 0.255288 +trainer/policy/normal/log_std Mean 0.996907 +trainer/policy/normal/log_std Std 0.297186 +trainer/policy/normal/log_std Max 1.71358 +trainer/policy/normal/log_std Min -1.36536 +eval/num steps total 609133 +eval/num paths total 611 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.297202 +eval/Actions Std 0.873986 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75557 +time/logging (s) 0.00408864 +time/sampling batch (s) 0.275919 +time/saving (s) 0.00388135 +time/training (s) 7.82966 +time/epoch (s) 10.8691 +time/total (s) 6338.06 +Epoch -390 +---------------------------------- --------------- +2022-05-10 14:56:32.979198 PDT | [1] Epoch -389 finished +---------------------------------- --------------- +epoch -389 +replay_buffer/size 999033 +trainer/num train calls 612000 +trainer/Policy Loss -19.8418 +trainer/Log Pis Mean 24.0061 +trainer/Log Pis Std 12.6333 +trainer/Log Pis Max 72.0053 +trainer/Log Pis Min -12.0608 +trainer/policy/mean Mean -0.0462114 +trainer/policy/mean Std 0.901126 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80105 +trainer/policy/normal/std Std 0.668277 +trainer/policy/normal/std Max 5.64252 +trainer/policy/normal/std Min 0.299229 +trainer/policy/normal/log_std Mean 0.988008 +trainer/policy/normal/log_std Std 0.329483 +trainer/policy/normal/log_std Max 1.73033 +trainer/policy/normal/log_std Min -1.20655 +eval/num steps total 610133 +eval/num paths total 612 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.087638 +eval/Actions Std 0.911385 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66334 +time/logging (s) 0.00370266 +time/sampling batch (s) 0.275829 +time/saving (s) 0.00353426 +time/training (s) 7.24935 +time/epoch (s) 10.1958 +time/total (s) 6348.26 +Epoch -389 +---------------------------------- --------------- +2022-05-10 14:56:43.867753 PDT | [1] Epoch -388 finished +---------------------------------- --------------- +epoch -388 +replay_buffer/size 999033 +trainer/num train calls 613000 +trainer/Policy Loss -19.7627 +trainer/Log Pis Mean 23.93 +trainer/Log Pis Std 12.6324 +trainer/Log Pis Max 64.9186 +trainer/Log Pis Min -7.74793 +trainer/policy/mean Mean -0.0244852 +trainer/policy/mean Std 0.905151 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7542 +trainer/policy/normal/std Std 0.63274 +trainer/policy/normal/std Max 5.81621 +trainer/policy/normal/std Min 0.247687 +trainer/policy/normal/log_std Mean 0.974956 +trainer/policy/normal/log_std Std 0.312242 +trainer/policy/normal/log_std Max 1.76065 +trainer/policy/normal/log_std Min -1.39559 +eval/num steps total 611133 +eval/num paths total 613 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.247884 +eval/Actions Std 0.919702 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6165 +time/logging (s) 0.00372319 +time/sampling batch (s) 1.02874 +time/saving (s) 0.00342633 +time/training (s) 7.21309 +time/epoch (s) 10.8655 +time/total (s) 6359.13 +Epoch -388 +---------------------------------- --------------- +2022-05-10 14:56:53.817268 PDT | [1] Epoch -387 finished +---------------------------------- --------------- +epoch -387 +replay_buffer/size 999033 +trainer/num train calls 614000 +trainer/Policy Loss -19.344 +trainer/Log Pis Mean 25.9271 +trainer/Log Pis Std 13.3038 +trainer/Log Pis Max 79.2492 +trainer/Log Pis Min -8.69427 +trainer/policy/mean Mean -0.0571699 +trainer/policy/mean Std 0.909821 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85132 +trainer/policy/normal/std Std 0.646043 +trainer/policy/normal/std Max 6.49044 +trainer/policy/normal/std Min 0.320295 +trainer/policy/normal/log_std Mean 1.01121 +trainer/policy/normal/log_std Std 0.304316 +trainer/policy/normal/log_std Max 1.87033 +trainer/policy/normal/log_std Min -1.13851 +eval/num steps total 612133 +eval/num paths total 614 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.093233 +eval/Actions Std 0.916689 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47315 +time/logging (s) 0.00390455 +time/sampling batch (s) 0.278281 +time/saving (s) 0.00354188 +time/training (s) 7.16761 +time/epoch (s) 9.92648 +time/total (s) 6369.06 +Epoch -387 +---------------------------------- --------------- +2022-05-10 14:57:04.063921 PDT | [1] Epoch -386 finished +---------------------------------- --------------- +epoch -386 +replay_buffer/size 999033 +trainer/num train calls 615000 +trainer/Policy Loss -19.3622 +trainer/Log Pis Mean 24.2703 +trainer/Log Pis Std 13.7741 +trainer/Log Pis Max 64.5497 +trainer/Log Pis Min -9.1963 +trainer/policy/mean Mean -0.018767 +trainer/policy/mean Std 0.904343 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.73478 +trainer/policy/normal/std Std 0.652158 +trainer/policy/normal/std Max 5.49943 +trainer/policy/normal/std Min 0.302034 +trainer/policy/normal/log_std Mean 0.964263 +trainer/policy/normal/log_std Std 0.328021 +trainer/policy/normal/log_std Max 1.70464 +trainer/policy/normal/log_std Min -1.19722 +eval/num steps total 613133 +eval/num paths total 615 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.494388 +eval/Actions Std 0.71438 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73667 +time/logging (s) 0.00373744 +time/sampling batch (s) 0.277636 +time/saving (s) 0.00341809 +time/training (s) 7.20186 +time/epoch (s) 10.2233 +time/total (s) 6379.29 +Epoch -386 +---------------------------------- --------------- +2022-05-10 14:57:15.140494 PDT | [1] Epoch -385 finished +---------------------------------- --------------- +epoch -385 +replay_buffer/size 999033 +trainer/num train calls 616000 +trainer/Policy Loss -19.2693 +trainer/Log Pis Mean 24.7508 +trainer/Log Pis Std 13.4817 +trainer/Log Pis Max 64.9115 +trainer/Log Pis Min -7.73053 +trainer/policy/mean Mean -0.0248367 +trainer/policy/mean Std 0.905556 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81005 +trainer/policy/normal/std Std 0.651782 +trainer/policy/normal/std Max 6.04837 +trainer/policy/normal/std Min 0.287157 +trainer/policy/normal/log_std Mean 0.993285 +trainer/policy/normal/log_std Std 0.321701 +trainer/policy/normal/log_std Max 1.79979 +trainer/policy/normal/log_std Min -1.24772 +eval/num steps total 614133 +eval/num paths total 616 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.234497 +eval/Actions Std 0.753694 +eval/Actions Max 0.999987 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61476 +time/logging (s) 0.00412498 +time/sampling batch (s) 0.277905 +time/saving (s) 0.00391316 +time/training (s) 8.15288 +time/epoch (s) 11.0536 +time/total (s) 6390.34 +Epoch -385 +---------------------------------- --------------- +2022-05-10 14:57:25.709029 PDT | [1] Epoch -384 finished +---------------------------------- --------------- +epoch -384 +replay_buffer/size 999033 +trainer/num train calls 617000 +trainer/Policy Loss -20.9438 +trainer/Log Pis Mean 24.7744 +trainer/Log Pis Std 13.2468 +trainer/Log Pis Max 68.4838 +trainer/Log Pis Min -8.00749 +trainer/policy/mean Mean -0.0427743 +trainer/policy/mean Std 0.909712 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.99997 +trainer/policy/normal/std Mean 2.81029 +trainer/policy/normal/std Std 0.648206 +trainer/policy/normal/std Max 5.36011 +trainer/policy/normal/std Min 0.28532 +trainer/policy/normal/log_std Mean 0.993495 +trainer/policy/normal/log_std Std 0.321239 +trainer/policy/normal/log_std Max 1.67898 +trainer/policy/normal/log_std Min -1.25415 +eval/num steps total 615133 +eval/num paths total 617 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.095142 +eval/Actions Std 0.920766 +eval/Actions Max 0.999988 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69465 +time/logging (s) 0.00380575 +time/sampling batch (s) 0.279298 +time/saving (s) 0.00354333 +time/training (s) 7.56336 +time/epoch (s) 10.5446 +time/total (s) 6400.89 +Epoch -384 +---------------------------------- --------------- +2022-05-10 14:57:37.474997 PDT | [1] Epoch -383 finished +---------------------------------- --------------- +epoch -383 +replay_buffer/size 999033 +trainer/num train calls 618000 +trainer/Policy Loss -18.8607 +trainer/Log Pis Mean 24.1766 +trainer/Log Pis Std 12.9703 +trainer/Log Pis Max 74.3843 +trainer/Log Pis Min -6.23534 +trainer/policy/mean Mean -0.0310264 +trainer/policy/mean Std 0.907657 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85626 +trainer/policy/normal/std Std 0.666979 +trainer/policy/normal/std Max 6.72363 +trainer/policy/normal/std Min 0.249111 +trainer/policy/normal/log_std Mean 1.01093 +trainer/policy/normal/log_std Std 0.313348 +trainer/policy/normal/log_std Max 1.90563 +trainer/policy/normal/log_std Min -1.38986 +eval/num steps total 616133 +eval/num paths total 618 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0391044 +eval/Actions Std 0.906769 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63463 +time/logging (s) 0.00377061 +time/sampling batch (s) 0.529144 +time/saving (s) 0.003548 +time/training (s) 8.57148 +time/epoch (s) 11.7426 +time/total (s) 6412.63 +Epoch -383 +---------------------------------- --------------- +2022-05-10 14:57:47.610300 PDT | [1] Epoch -382 finished +---------------------------------- --------------- +epoch -382 +replay_buffer/size 999033 +trainer/num train calls 619000 +trainer/Policy Loss -21.5679 +trainer/Log Pis Mean 24.7939 +trainer/Log Pis Std 13.2898 +trainer/Log Pis Max 70.3897 +trainer/Log Pis Min -8.43824 +trainer/policy/mean Mean -0.0399502 +trainer/policy/mean Std 0.906073 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77861 +trainer/policy/normal/std Std 0.623758 +trainer/policy/normal/std Max 6.02262 +trainer/policy/normal/std Min 0.333674 +trainer/policy/normal/log_std Mean 0.985774 +trainer/policy/normal/log_std Std 0.303958 +trainer/policy/normal/log_std Max 1.79552 +trainer/policy/normal/log_std Min -1.09759 +eval/num steps total 617133 +eval/num paths total 619 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0242896 +eval/Actions Std 0.908204 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43016 +time/logging (s) 0.00385791 +time/sampling batch (s) 0.277594 +time/saving (s) 0.00345355 +time/training (s) 7.39712 +time/epoch (s) 10.1122 +time/total (s) 6422.75 +Epoch -382 +---------------------------------- --------------- +2022-05-10 14:57:57.981378 PDT | [1] Epoch -381 finished +---------------------------------- --------------- +epoch -381 +replay_buffer/size 999033 +trainer/num train calls 620000 +trainer/Policy Loss -19.8454 +trainer/Log Pis Mean 24.5278 +trainer/Log Pis Std 12.9409 +trainer/Log Pis Max 63.8622 +trainer/Log Pis Min -8.73118 +trainer/policy/mean Mean -0.0177436 +trainer/policy/mean Std 0.905474 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75827 +trainer/policy/normal/std Std 0.643518 +trainer/policy/normal/std Max 6.25791 +trainer/policy/normal/std Min 0.286606 +trainer/policy/normal/log_std Mean 0.975711 +trainer/policy/normal/log_std Std 0.314439 +trainer/policy/normal/log_std Max 1.83385 +trainer/policy/normal/log_std Min -1.24965 +eval/num steps total 618133 +eval/num paths total 620 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108732 +eval/Actions Std 0.923062 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56011 +time/logging (s) 0.00367167 +time/sampling batch (s) 0.276243 +time/saving (s) 0.0034497 +time/training (s) 7.50446 +time/epoch (s) 10.3479 +time/total (s) 6433.1 +Epoch -381 +---------------------------------- --------------- +2022-05-10 14:58:07.793869 PDT | [1] Epoch -380 finished +---------------------------------- --------------- +epoch -380 +replay_buffer/size 999033 +trainer/num train calls 621000 +trainer/Policy Loss -18.499 +trainer/Log Pis Mean 22.5122 +trainer/Log Pis Std 13.2595 +trainer/Log Pis Max 62.9199 +trainer/Log Pis Min -9.62182 +trainer/policy/mean Mean -0.0280004 +trainer/policy/mean Std 0.901906 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.75132 +trainer/policy/normal/std Std 0.66304 +trainer/policy/normal/std Max 6.45966 +trainer/policy/normal/std Min 0.255003 +trainer/policy/normal/log_std Mean 0.969499 +trainer/policy/normal/log_std Std 0.331154 +trainer/policy/normal/log_std Max 1.86558 +trainer/policy/normal/log_std Min -1.36648 +eval/num steps total 619133 +eval/num paths total 621 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.313321 +eval/Actions Std 0.857443 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56649 +time/logging (s) 0.00411488 +time/sampling batch (s) 0.273877 +time/saving (s) 0.00395423 +time/training (s) 6.94158 +time/epoch (s) 9.79002 +time/total (s) 6442.89 +Epoch -380 +---------------------------------- --------------- +2022-05-10 14:58:18.493681 PDT | [1] Epoch -379 finished +---------------------------------- --------------- +epoch -379 +replay_buffer/size 999033 +trainer/num train calls 622000 +trainer/Policy Loss -20.7068 +trainer/Log Pis Mean 24.5303 +trainer/Log Pis Std 13.3261 +trainer/Log Pis Max 65.3771 +trainer/Log Pis Min -7.72919 +trainer/policy/mean Mean -0.038987 +trainer/policy/mean Std 0.905648 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8264 +trainer/policy/normal/std Std 0.651053 +trainer/policy/normal/std Max 6.20704 +trainer/policy/normal/std Min 0.307064 +trainer/policy/normal/log_std Mean 1.00043 +trainer/policy/normal/log_std Std 0.314841 +trainer/policy/normal/log_std Max 1.82568 +trainer/policy/normal/log_std Min -1.1807 +eval/num steps total 620133 +eval/num paths total 622 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00497793 +eval/Actions Std 0.903483 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63484 +time/logging (s) 0.00372799 +time/sampling batch (s) 0.274395 +time/saving (s) 0.00357379 +time/training (s) 7.75947 +time/epoch (s) 10.676 +time/total (s) 6453.57 +Epoch -379 +---------------------------------- --------------- +2022-05-10 14:58:28.646138 PDT | [1] Epoch -378 finished +---------------------------------- --------------- +epoch -378 +replay_buffer/size 999033 +trainer/num train calls 623000 +trainer/Policy Loss -19.3536 +trainer/Log Pis Mean 24.37 +trainer/Log Pis Std 13.7672 +trainer/Log Pis Max 69.0161 +trainer/Log Pis Min -10.3133 +trainer/policy/mean Mean -0.0294041 +trainer/policy/mean Std 0.905757 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86333 +trainer/policy/normal/std Std 0.665584 +trainer/policy/normal/std Max 5.31595 +trainer/policy/normal/std Min 0.309957 +trainer/policy/normal/log_std Mean 1.01272 +trainer/policy/normal/log_std Std 0.317057 +trainer/policy/normal/log_std Max 1.67071 +trainer/policy/normal/log_std Min -1.17132 +eval/num steps total 621133 +eval/num paths total 623 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102069 +eval/Actions Std 0.895195 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52721 +time/logging (s) 0.00390417 +time/sampling batch (s) 0.529117 +time/saving (s) 0.00346541 +time/training (s) 7.06551 +time/epoch (s) 10.1292 +time/total (s) 6463.71 +Epoch -378 +---------------------------------- --------------- +2022-05-10 14:58:38.799195 PDT | [1] Epoch -377 finished +---------------------------------- --------------- +epoch -377 +replay_buffer/size 999033 +trainer/num train calls 624000 +trainer/Policy Loss -19.5576 +trainer/Log Pis Mean 24.9261 +trainer/Log Pis Std 13.5018 +trainer/Log Pis Max 59.4948 +trainer/Log Pis Min -6.85599 +trainer/policy/mean Mean -0.0423865 +trainer/policy/mean Std 0.905492 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.88733 +trainer/policy/normal/std Std 0.662894 +trainer/policy/normal/std Max 6.50298 +trainer/policy/normal/std Min 0.291211 +trainer/policy/normal/log_std Mean 1.02308 +trainer/policy/normal/log_std Std 0.307136 +trainer/policy/normal/log_std Max 1.87226 +trainer/policy/normal/log_std Min -1.23371 +eval/num steps total 622133 +eval/num paths total 624 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12496 +eval/Actions Std 0.882004 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5904 +time/logging (s) 0.00370901 +time/sampling batch (s) 0.277811 +time/saving (s) 0.0035371 +time/training (s) 7.25425 +time/epoch (s) 10.1297 +time/total (s) 6473.84 +Epoch -377 +---------------------------------- --------------- +2022-05-10 14:58:48.584134 PDT | [1] Epoch -376 finished +---------------------------------- --------------- +epoch -376 +replay_buffer/size 999033 +trainer/num train calls 625000 +trainer/Policy Loss -19.7977 +trainer/Log Pis Mean 24.3927 +trainer/Log Pis Std 13.3547 +trainer/Log Pis Max 66.3435 +trainer/Log Pis Min -5.77734 +trainer/policy/mean Mean -0.0305633 +trainer/policy/mean Std 0.902276 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82055 +trainer/policy/normal/std Std 0.679131 +trainer/policy/normal/std Max 5.90565 +trainer/policy/normal/std Min 0.325819 +trainer/policy/normal/log_std Mean 0.994165 +trainer/policy/normal/log_std Std 0.332103 +trainer/policy/normal/log_std Max 1.77591 +trainer/policy/normal/log_std Min -1.12141 +eval/num steps total 623133 +eval/num paths total 625 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0798089 +eval/Actions Std 0.908178 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50771 +time/logging (s) 0.00381396 +time/sampling batch (s) 0.280806 +time/saving (s) 0.00345368 +time/training (s) 6.96594 +time/epoch (s) 9.76173 +time/total (s) 6483.6 +Epoch -376 +---------------------------------- --------------- +2022-05-10 14:58:59.892272 PDT | [1] Epoch -375 finished +---------------------------------- --------------- +epoch -375 +replay_buffer/size 999033 +trainer/num train calls 626000 +trainer/Policy Loss -20.6259 +trainer/Log Pis Mean 24.5731 +trainer/Log Pis Std 12.9863 +trainer/Log Pis Max 76.7093 +trainer/Log Pis Min -9.65004 +trainer/policy/mean Mean -0.0529574 +trainer/policy/mean Std 0.907968 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.78907 +trainer/policy/normal/std Std 0.646991 +trainer/policy/normal/std Max 6.0579 +trainer/policy/normal/std Min 0.305041 +trainer/policy/normal/log_std Mean 0.987296 +trainer/policy/normal/log_std Std 0.311532 +trainer/policy/normal/log_std Max 1.80136 +trainer/policy/normal/log_std Min -1.18731 +eval/num steps total 624133 +eval/num paths total 626 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.340414 +eval/Actions Std 0.854923 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63114 +time/logging (s) 0.00378614 +time/sampling batch (s) 0.53373 +time/saving (s) 0.00341951 +time/training (s) 8.11234 +time/epoch (s) 11.2844 +time/total (s) 6494.89 +Epoch -375 +---------------------------------- --------------- +2022-05-10 14:59:10.029855 PDT | [1] Epoch -374 finished +---------------------------------- --------------- +epoch -374 +replay_buffer/size 999033 +trainer/num train calls 627000 +trainer/Policy Loss -19.5042 +trainer/Log Pis Mean 24.9989 +trainer/Log Pis Std 13.6329 +trainer/Log Pis Max 67.0994 +trainer/Log Pis Min -10.9099 +trainer/policy/mean Mean -0.0364493 +trainer/policy/mean Std 0.905219 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80799 +trainer/policy/normal/std Std 0.654073 +trainer/policy/normal/std Max 5.41329 +trainer/policy/normal/std Min 0.254935 +trainer/policy/normal/log_std Mean 0.993492 +trainer/policy/normal/log_std Std 0.315234 +trainer/policy/normal/log_std Max 1.68886 +trainer/policy/normal/log_std Min -1.36675 +eval/num steps total 625133 +eval/num paths total 627 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216468 +eval/Actions Std 0.796167 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55028 +time/logging (s) 0.00406895 +time/sampling batch (s) 0.281461 +time/saving (s) 0.00397988 +time/training (s) 7.27441 +time/epoch (s) 10.1142 +time/total (s) 6505.01 +Epoch -374 +---------------------------------- --------------- +2022-05-10 14:59:20.133847 PDT | [1] Epoch -373 finished +---------------------------------- --------------- +epoch -373 +replay_buffer/size 999033 +trainer/num train calls 628000 +trainer/Policy Loss -19.246 +trainer/Log Pis Mean 24.7692 +trainer/Log Pis Std 13.905 +trainer/Log Pis Max 70.3604 +trainer/Log Pis Min -7.00957 +trainer/policy/mean Mean -0.0279345 +trainer/policy/mean Std 0.906587 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85429 +trainer/policy/normal/std Std 0.659565 +trainer/policy/normal/std Max 7.09191 +trainer/policy/normal/std Min 0.250828 +trainer/policy/normal/log_std Mean 1.01032 +trainer/policy/normal/log_std Std 0.313512 +trainer/policy/normal/log_std Max 1.95895 +trainer/policy/normal/log_std Min -1.38299 +eval/num steps total 626133 +eval/num paths total 628 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0730448 +eval/Actions Std 0.914206 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60796 +time/logging (s) 0.00441054 +time/sampling batch (s) 0.280367 +time/saving (s) 0.00360597 +time/training (s) 7.18421 +time/epoch (s) 10.0805 +time/total (s) 6515.09 +Epoch -373 +---------------------------------- --------------- +2022-05-10 14:59:30.871964 PDT | [1] Epoch -372 finished +---------------------------------- --------------- +epoch -372 +replay_buffer/size 999033 +trainer/num train calls 629000 +trainer/Policy Loss -20.7463 +trainer/Log Pis Mean 24.0641 +trainer/Log Pis Std 12.4262 +trainer/Log Pis Max 75.2637 +trainer/Log Pis Min -5.80733 +trainer/policy/mean Mean -0.0404654 +trainer/policy/mean Std 0.903904 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.76465 +trainer/policy/normal/std Std 0.637736 +trainer/policy/normal/std Max 5.9604 +trainer/policy/normal/std Min 0.38947 +trainer/policy/normal/log_std Mean 0.978486 +trainer/policy/normal/log_std Std 0.313424 +trainer/policy/normal/log_std Max 1.78514 +trainer/policy/normal/log_std Min -0.942968 +eval/num steps total 627133 +eval/num paths total 629 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0156493 +eval/Actions Std 0.911587 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4659 +time/logging (s) 0.0036842 +time/sampling batch (s) 0.529302 +time/saving (s) 0.00342524 +time/training (s) 7.71162 +time/epoch (s) 10.7139 +time/total (s) 6525.81 +Epoch -372 +---------------------------------- --------------- +2022-05-10 14:59:42.335136 PDT | [1] Epoch -371 finished +---------------------------------- --------------- +epoch -371 +replay_buffer/size 999033 +trainer/num train calls 630000 +trainer/Policy Loss -20.4741 +trainer/Log Pis Mean 24.0813 +trainer/Log Pis Std 13.4514 +trainer/Log Pis Max 72.1242 +trainer/Log Pis Min -8.10891 +trainer/policy/mean Mean -0.0270967 +trainer/policy/mean Std 0.904774 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78564 +trainer/policy/normal/std Std 0.651517 +trainer/policy/normal/std Max 6.37193 +trainer/policy/normal/std Min 0.324013 +trainer/policy/normal/log_std Mean 0.986099 +trainer/policy/normal/log_std Std 0.310774 +trainer/policy/normal/log_std Max 1.8519 +trainer/policy/normal/log_std Min -1.12697 +eval/num steps total 628133 +eval/num paths total 630 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0488848 +eval/Actions Std 0.901598 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47524 +time/logging (s) 0.00386556 +time/sampling batch (s) 0.528074 +time/saving (s) 0.00345929 +time/training (s) 8.42964 +time/epoch (s) 11.4403 +time/total (s) 6537.25 +Epoch -371 +---------------------------------- --------------- +2022-05-10 14:59:52.483276 PDT | [1] Epoch -370 finished +---------------------------------- --------------- +epoch -370 +replay_buffer/size 999033 +trainer/num train calls 631000 +trainer/Policy Loss -20.106 +trainer/Log Pis Mean 24.5145 +trainer/Log Pis Std 12.7922 +trainer/Log Pis Max 66.943 +trainer/Log Pis Min -9.52688 +trainer/policy/mean Mean -0.0424478 +trainer/policy/mean Std 0.905316 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84538 +trainer/policy/normal/std Std 0.643862 +trainer/policy/normal/std Max 7.01099 +trainer/policy/normal/std Min 0.278581 +trainer/policy/normal/log_std Mean 1.01048 +trainer/policy/normal/log_std Std 0.29671 +trainer/policy/normal/log_std Max 1.94748 +trainer/policy/normal/log_std Min -1.27805 +eval/num steps total 629133 +eval/num paths total 631 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.211581 +eval/Actions Std 0.85496 +eval/Actions Max 0.999981 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69964 +time/logging (s) 0.00375957 +time/sampling batch (s) 0.275359 +time/saving (s) 0.00343776 +time/training (s) 7.14287 +time/epoch (s) 10.1251 +time/total (s) 6547.38 +Epoch -370 +---------------------------------- --------------- +2022-05-10 15:00:02.905925 PDT | [1] Epoch -369 finished +---------------------------------- --------------- +epoch -369 +replay_buffer/size 999033 +trainer/num train calls 632000 +trainer/Policy Loss -18.5218 +trainer/Log Pis Mean 24.2589 +trainer/Log Pis Std 12.9963 +trainer/Log Pis Max 72.663 +trainer/Log Pis Min -6.61746 +trainer/policy/mean Mean -0.0103422 +trainer/policy/mean Std 0.910529 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999967 +trainer/policy/normal/std Mean 2.84691 +trainer/policy/normal/std Std 0.646253 +trainer/policy/normal/std Max 6.54176 +trainer/policy/normal/std Min 0.278179 +trainer/policy/normal/log_std Mean 1.01023 +trainer/policy/normal/log_std Std 0.300138 +trainer/policy/normal/log_std Max 1.87821 +trainer/policy/normal/log_std Min -1.27949 +eval/num steps total 630133 +eval/num paths total 632 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.162414 +eval/Actions Std 0.87142 +eval/Actions Max 0.999971 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71214 +time/logging (s) 0.00410159 +time/sampling batch (s) 0.28495 +time/saving (s) 0.00398581 +time/training (s) 7.39428 +time/epoch (s) 10.3995 +time/total (s) 6557.78 +Epoch -369 +---------------------------------- --------------- +2022-05-10 15:00:12.950880 PDT | [1] Epoch -368 finished +---------------------------------- --------------- +epoch -368 +replay_buffer/size 999033 +trainer/num train calls 633000 +trainer/Policy Loss -20.1982 +trainer/Log Pis Mean 24.1803 +trainer/Log Pis Std 13.1245 +trainer/Log Pis Max 73.389 +trainer/Log Pis Min -8.31257 +trainer/policy/mean Mean -0.0427156 +trainer/policy/mean Std 0.903424 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80432 +trainer/policy/normal/std Std 0.657619 +trainer/policy/normal/std Max 5.7507 +trainer/policy/normal/std Min 0.277418 +trainer/policy/normal/log_std Mean 0.990447 +trainer/policy/normal/log_std Std 0.324544 +trainer/policy/normal/log_std Max 1.74932 +trainer/policy/normal/log_std Min -1.28223 +eval/num steps total 631133 +eval/num paths total 633 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.48438 +eval/Actions Std 0.867784 +eval/Actions Max 0.99996 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73302 +time/logging (s) 0.00376262 +time/sampling batch (s) 0.281102 +time/saving (s) 0.00352852 +time/training (s) 6.9993 +time/epoch (s) 10.0207 +time/total (s) 6567.81 +Epoch -368 +---------------------------------- --------------- +2022-05-10 15:00:23.617520 PDT | [1] Epoch -367 finished +---------------------------------- --------------- +epoch -367 +replay_buffer/size 999033 +trainer/num train calls 634000 +trainer/Policy Loss -19.1876 +trainer/Log Pis Mean 25.0726 +trainer/Log Pis Std 13.4743 +trainer/Log Pis Max 72.9974 +trainer/Log Pis Min -6.83509 +trainer/policy/mean Mean -0.0477929 +trainer/policy/mean Std 0.908831 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7986 +trainer/policy/normal/std Std 0.637308 +trainer/policy/normal/std Max 5.477 +trainer/policy/normal/std Min 0.233061 +trainer/policy/normal/log_std Mean 0.992174 +trainer/policy/normal/log_std Std 0.306365 +trainer/policy/normal/log_std Max 1.70056 +trainer/policy/normal/log_std Min -1.45646 +eval/num steps total 632133 +eval/num paths total 634 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.14015 +eval/Actions Std 0.909718 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71739 +time/logging (s) 0.0037443 +time/sampling batch (s) 0.285637 +time/saving (s) 0.00348163 +time/training (s) 7.63249 +time/epoch (s) 10.6427 +time/total (s) 6578.45 +Epoch -367 +---------------------------------- --------------- +2022-05-10 15:00:33.199172 PDT | [1] Epoch -366 finished +---------------------------------- --------------- +epoch -366 +replay_buffer/size 999033 +trainer/num train calls 635000 +trainer/Policy Loss -18.7735 +trainer/Log Pis Mean 24.2725 +trainer/Log Pis Std 12.9551 +trainer/Log Pis Max 67.02 +trainer/Log Pis Min -7.94376 +trainer/policy/mean Mean -0.0261866 +trainer/policy/mean Std 0.91047 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77311 +trainer/policy/normal/std Std 0.649069 +trainer/policy/normal/std Max 5.53173 +trainer/policy/normal/std Min 0.227591 +trainer/policy/normal/log_std Mean 0.978816 +trainer/policy/normal/log_std Std 0.327901 +trainer/policy/normal/log_std Max 1.7105 +trainer/policy/normal/log_std Min -1.48021 +eval/num steps total 633133 +eval/num paths total 635 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0657034 +eval/Actions Std 0.91429 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52222 +time/logging (s) 0.00378994 +time/sampling batch (s) 0.53029 +time/saving (s) 0.00342838 +time/training (s) 6.49854 +time/epoch (s) 9.55827 +time/total (s) 6588.01 +Epoch -366 +---------------------------------- --------------- +2022-05-10 15:00:43.103696 PDT | [1] Epoch -365 finished +---------------------------------- --------------- +epoch -365 +replay_buffer/size 999033 +trainer/num train calls 636000 +trainer/Policy Loss -20.8247 +trainer/Log Pis Mean 24.1069 +trainer/Log Pis Std 13.3135 +trainer/Log Pis Max 65.2121 +trainer/Log Pis Min -9.4964 +trainer/policy/mean Mean -0.0374709 +trainer/policy/mean Std 0.910602 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77339 +trainer/policy/normal/std Std 0.63105 +trainer/policy/normal/std Max 6.61042 +trainer/policy/normal/std Min 0.342134 +trainer/policy/normal/log_std Mean 0.983757 +trainer/policy/normal/log_std Std 0.301983 +trainer/policy/normal/log_std Max 1.88865 +trainer/policy/normal/log_std Min -1.07255 +eval/num steps total 634133 +eval/num paths total 636 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108778 +eval/Actions Std 0.886368 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42198 +time/logging (s) 0.00375339 +time/sampling batch (s) 0.526423 +time/saving (s) 0.00343155 +time/training (s) 6.92585 +time/epoch (s) 9.88144 +time/total (s) 6597.9 +Epoch -365 +---------------------------------- --------------- +2022-05-10 15:00:53.934712 PDT | [1] Epoch -364 finished +---------------------------------- --------------- +epoch -364 +replay_buffer/size 999033 +trainer/num train calls 637000 +trainer/Policy Loss -20.3133 +trainer/Log Pis Mean 24.6243 +trainer/Log Pis Std 12.6173 +trainer/Log Pis Max 63.8773 +trainer/Log Pis Min -7.85833 +trainer/policy/mean Mean -0.00892652 +trainer/policy/mean Std 0.909355 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80634 +trainer/policy/normal/std Std 0.648114 +trainer/policy/normal/std Max 6.47418 +trainer/policy/normal/std Min 0.277918 +trainer/policy/normal/log_std Mean 0.993838 +trainer/policy/normal/log_std Std 0.310926 +trainer/policy/normal/log_std Max 1.86782 +trainer/policy/normal/log_std Min -1.28043 +eval/num steps total 635133 +eval/num paths total 637 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0639538 +eval/Actions Std 0.907075 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31673 +time/logging (s) 0.00418612 +time/sampling batch (s) 1.02805 +time/saving (s) 0.00385762 +time/training (s) 7.45536 +time/epoch (s) 10.8082 +time/total (s) 6608.71 +Epoch -364 +---------------------------------- --------------- +2022-05-10 15:01:04.706461 PDT | [1] Epoch -363 finished +---------------------------------- --------------- +epoch -363 +replay_buffer/size 999033 +trainer/num train calls 638000 +trainer/Policy Loss -19.3615 +trainer/Log Pis Mean 25.2995 +trainer/Log Pis Std 13.9937 +trainer/Log Pis Max 69.1938 +trainer/Log Pis Min -8.92935 +trainer/policy/mean Mean -0.0212271 +trainer/policy/mean Std 0.907969 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81088 +trainer/policy/normal/std Std 0.664153 +trainer/policy/normal/std Max 6.25278 +trainer/policy/normal/std Min 0.271427 +trainer/policy/normal/log_std Mean 0.992261 +trainer/policy/normal/log_std Std 0.328285 +trainer/policy/normal/log_std Max 1.83303 +trainer/policy/normal/log_std Min -1.30406 +eval/num steps total 636133 +eval/num paths total 638 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.479573 +eval/Actions Std 0.81985 +eval/Actions Max 0.999995 +eval/Actions Min -0.999963 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60886 +time/logging (s) 0.00403127 +time/sampling batch (s) 0.531238 +time/saving (s) 0.00379764 +time/training (s) 7.59975 +time/epoch (s) 10.7477 +time/total (s) 6619.46 +Epoch -363 +---------------------------------- --------------- +2022-05-10 15:01:15.132357 PDT | [1] Epoch -362 finished +---------------------------------- --------------- +epoch -362 +replay_buffer/size 999033 +trainer/num train calls 639000 +trainer/Policy Loss -19.4406 +trainer/Log Pis Mean 23.7676 +trainer/Log Pis Std 13.5385 +trainer/Log Pis Max 70.2479 +trainer/Log Pis Min -2.90557 +trainer/policy/mean Mean -0.0519937 +trainer/policy/mean Std 0.906204 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81238 +trainer/policy/normal/std Std 0.649674 +trainer/policy/normal/std Max 5.97404 +trainer/policy/normal/std Min 0.311631 +trainer/policy/normal/log_std Mean 0.994769 +trainer/policy/normal/log_std Std 0.318263 +trainer/policy/normal/log_std Max 1.78742 +trainer/policy/normal/log_std Min -1.16594 +eval/num steps total 637133 +eval/num paths total 639 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108929 +eval/Actions Std 0.903278 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59967 +time/logging (s) 0.00383842 +time/sampling batch (s) 0.530225 +time/saving (s) 0.0034194 +time/training (s) 7.26392 +time/epoch (s) 10.4011 +time/total (s) 6629.86 +Epoch -362 +---------------------------------- --------------- +2022-05-10 15:01:25.055765 PDT | [1] Epoch -361 finished +---------------------------------- --------------- +epoch -361 +replay_buffer/size 999033 +trainer/num train calls 640000 +trainer/Policy Loss -21.1985 +trainer/Log Pis Mean 24.7048 +trainer/Log Pis Std 12.7757 +trainer/Log Pis Max 69.6567 +trainer/Log Pis Min -4.52097 +trainer/policy/mean Mean -0.0329711 +trainer/policy/mean Std 0.907946 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.78996 +trainer/policy/normal/std Std 0.639365 +trainer/policy/normal/std Max 6.6907 +trainer/policy/normal/std Min 0.349127 +trainer/policy/normal/log_std Mean 0.989114 +trainer/policy/normal/log_std Std 0.304871 +trainer/policy/normal/log_std Max 1.90072 +trainer/policy/normal/log_std Min -1.05232 +eval/num steps total 638133 +eval/num paths total 640 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0965613 +eval/Actions Std 0.919222 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50831 +time/logging (s) 0.00428537 +time/sampling batch (s) 0.277405 +time/saving (s) 0.0041056 +time/training (s) 7.10578 +time/epoch (s) 9.89988 +time/total (s) 6639.77 +Epoch -361 +---------------------------------- --------------- +2022-05-10 15:01:34.429138 PDT | [1] Epoch -360 finished +---------------------------------- --------------- +epoch -360 +replay_buffer/size 999033 +trainer/num train calls 641000 +trainer/Policy Loss -19.0646 +trainer/Log Pis Mean 24.9176 +trainer/Log Pis Std 12.8683 +trainer/Log Pis Max 67.385 +trainer/Log Pis Min -9.63337 +trainer/policy/mean Mean -0.0387904 +trainer/policy/mean Std 0.908867 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83482 +trainer/policy/normal/std Std 0.643571 +trainer/policy/normal/std Max 5.56986 +trainer/policy/normal/std Min 0.295114 +trainer/policy/normal/log_std Mean 1.00457 +trainer/policy/normal/log_std Std 0.309637 +trainer/policy/normal/log_std Max 1.71737 +trainer/policy/normal/log_std Min -1.2204 +eval/num steps total 639133 +eval/num paths total 641 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0141382 +eval/Actions Std 0.916174 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58827 +time/logging (s) 0.00421228 +time/sampling batch (s) 0.293916 +time/saving (s) 0.00419237 +time/training (s) 6.4576 +time/epoch (s) 9.34819 +time/total (s) 6649.12 +Epoch -360 +---------------------------------- --------------- +2022-05-10 15:01:44.702912 PDT | [1] Epoch -359 finished +---------------------------------- --------------- +epoch -359 +replay_buffer/size 999033 +trainer/num train calls 642000 +trainer/Policy Loss -19.7815 +trainer/Log Pis Mean 24.7168 +trainer/Log Pis Std 14.1886 +trainer/Log Pis Max 84.161 +trainer/Log Pis Min -8.46553 +trainer/policy/mean Mean -0.0337107 +trainer/policy/mean Std 0.9092 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78899 +trainer/policy/normal/std Std 0.656652 +trainer/policy/normal/std Max 5.86679 +trainer/policy/normal/std Min 0.268872 +trainer/policy/normal/log_std Mean 0.985009 +trainer/policy/normal/log_std Std 0.323125 +trainer/policy/normal/log_std Max 1.76931 +trainer/policy/normal/log_std Min -1.31352 +eval/num steps total 640133 +eval/num paths total 642 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.225893 +eval/Actions Std 0.938376 +eval/Actions Max 0.999996 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63817 +time/logging (s) 0.00394356 +time/sampling batch (s) 0.559838 +time/saving (s) 0.00379931 +time/training (s) 7.04313 +time/epoch (s) 10.2489 +time/total (s) 6659.37 +Epoch -359 +---------------------------------- --------------- +2022-05-10 15:01:55.066521 PDT | [1] Epoch -358 finished +---------------------------------- --------------- +epoch -358 +replay_buffer/size 999033 +trainer/num train calls 643000 +trainer/Policy Loss -19.6214 +trainer/Log Pis Mean 23.8728 +trainer/Log Pis Std 12.7059 +trainer/Log Pis Max 61.5665 +trainer/Log Pis Min -10.4163 +trainer/policy/mean Mean -0.0256238 +trainer/policy/mean Std 0.903539 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78317 +trainer/policy/normal/std Std 0.674548 +trainer/policy/normal/std Max 6.92039 +trainer/policy/normal/std Min 0.270253 +trainer/policy/normal/log_std Mean 0.981698 +trainer/policy/normal/log_std Std 0.325952 +trainer/policy/normal/log_std Max 1.93447 +trainer/policy/normal/log_std Min -1.3084 +eval/num steps total 641133 +eval/num paths total 643 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00531935 +eval/Actions Std 0.978435 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67569 +time/logging (s) 0.00410013 +time/sampling batch (s) 0.331132 +time/saving (s) 0.0041764 +time/training (s) 7.32441 +time/epoch (s) 10.3395 +time/total (s) 6669.71 +Epoch -358 +---------------------------------- --------------- +2022-05-10 15:02:06.407337 PDT | [1] Epoch -357 finished +---------------------------------- --------------- +epoch -357 +replay_buffer/size 999033 +trainer/num train calls 644000 +trainer/Policy Loss -19.1049 +trainer/Log Pis Mean 24.686 +trainer/Log Pis Std 13.8998 +trainer/Log Pis Max 76.3789 +trainer/Log Pis Min -6.43368 +trainer/policy/mean Mean -0.0240644 +trainer/policy/mean Std 0.905983 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84102 +trainer/policy/normal/std Std 0.681206 +trainer/policy/normal/std Max 6.82153 +trainer/policy/normal/std Min 0.304616 +trainer/policy/normal/log_std Mean 1.00345 +trainer/policy/normal/log_std Std 0.319954 +trainer/policy/normal/log_std Max 1.92008 +trainer/policy/normal/log_std Min -1.1887 +eval/num steps total 642133 +eval/num paths total 644 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222484 +eval/Actions Std 0.872086 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80091 +time/logging (s) 0.00366855 +time/sampling batch (s) 0.277132 +time/saving (s) 0.00344393 +time/training (s) 8.23161 +time/epoch (s) 11.3168 +time/total (s) 6681.03 +Epoch -357 +---------------------------------- --------------- +2022-05-10 15:02:16.827870 PDT | [1] Epoch -356 finished +---------------------------------- --------------- +epoch -356 +replay_buffer/size 999033 +trainer/num train calls 645000 +trainer/Policy Loss -19.9241 +trainer/Log Pis Mean 24.1419 +trainer/Log Pis Std 13.4423 +trainer/Log Pis Max 77.0214 +trainer/Log Pis Min -6.66793 +trainer/policy/mean Mean -0.0194504 +trainer/policy/mean Std 0.906601 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80479 +trainer/policy/normal/std Std 0.658179 +trainer/policy/normal/std Max 5.84741 +trainer/policy/normal/std Min 0.292934 +trainer/policy/normal/log_std Mean 0.992911 +trainer/policy/normal/log_std Std 0.309493 +trainer/policy/normal/log_std Max 1.766 +trainer/policy/normal/log_std Min -1.22781 +eval/num steps total 643133 +eval/num paths total 645 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.164427 +eval/Actions Std 0.936755 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64156 +time/logging (s) 0.00426713 +time/sampling batch (s) 0.27554 +time/saving (s) 0.00399439 +time/training (s) 7.47249 +time/epoch (s) 10.3979 +time/total (s) 6691.44 +Epoch -356 +---------------------------------- --------------- +2022-05-10 15:02:29.038172 PDT | [1] Epoch -355 finished +---------------------------------- --------------- +epoch -355 +replay_buffer/size 999033 +trainer/num train calls 646000 +trainer/Policy Loss -19.5811 +trainer/Log Pis Mean 24.1471 +trainer/Log Pis Std 13.1446 +trainer/Log Pis Max 74.4475 +trainer/Log Pis Min -12.5572 +trainer/policy/mean Mean -0.0392813 +trainer/policy/mean Std 0.908246 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76655 +trainer/policy/normal/std Std 0.628966 +trainer/policy/normal/std Max 5.23086 +trainer/policy/normal/std Min 0.303044 +trainer/policy/normal/log_std Mean 0.979982 +trainer/policy/normal/log_std Std 0.310287 +trainer/policy/normal/log_std Max 1.65458 +trainer/policy/normal/log_std Min -1.19388 +eval/num steps total 644133 +eval/num paths total 646 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0560156 +eval/Actions Std 0.855365 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56054 +time/logging (s) 0.00376442 +time/sampling batch (s) 0.529547 +time/saving (s) 0.00372566 +time/training (s) 9.08828 +time/epoch (s) 12.1859 +time/total (s) 6703.62 +Epoch -355 +---------------------------------- --------------- +2022-05-10 15:02:40.018795 PDT | [1] Epoch -354 finished +---------------------------------- --------------- +epoch -354 +replay_buffer/size 999033 +trainer/num train calls 647000 +trainer/Policy Loss -19.6306 +trainer/Log Pis Mean 23.9568 +trainer/Log Pis Std 13.7501 +trainer/Log Pis Max 64.309 +trainer/Log Pis Min -7.31996 +trainer/policy/mean Mean -0.0304039 +trainer/policy/mean Std 0.907195 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82248 +trainer/policy/normal/std Std 0.64207 +trainer/policy/normal/std Max 5.98256 +trainer/policy/normal/std Min 0.30967 +trainer/policy/normal/log_std Mean 1.00178 +trainer/policy/normal/log_std Std 0.297883 +trainer/policy/normal/log_std Max 1.78885 +trainer/policy/normal/log_std Min -1.17225 +eval/num steps total 645133 +eval/num paths total 647 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.149931 +eval/Actions Std 0.867368 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.22569 +time/logging (s) 0.00366008 +time/sampling batch (s) 0.527675 +time/saving (s) 0.00338362 +time/training (s) 8.1968 +time/epoch (s) 10.9572 +time/total (s) 6714.58 +Epoch -354 +---------------------------------- --------------- +2022-05-10 15:02:50.879446 PDT | [1] Epoch -353 finished +---------------------------------- --------------- +epoch -353 +replay_buffer/size 999033 +trainer/num train calls 648000 +trainer/Policy Loss -19.4605 +trainer/Log Pis Mean 24.1324 +trainer/Log Pis Std 13.046 +trainer/Log Pis Max 69.4558 +trainer/Log Pis Min -4.43715 +trainer/policy/mean Mean -0.0414181 +trainer/policy/mean Std 0.904973 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83557 +trainer/policy/normal/std Std 0.665599 +trainer/policy/normal/std Max 5.39815 +trainer/policy/normal/std Min 0.273361 +trainer/policy/normal/log_std Mean 1.00174 +trainer/policy/normal/log_std Std 0.322072 +trainer/policy/normal/log_std Max 1.68606 +trainer/policy/normal/log_std Min -1.29696 +eval/num steps total 646133 +eval/num paths total 648 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116131 +eval/Actions Std 0.909727 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60414 +time/logging (s) 0.00379577 +time/sampling batch (s) 0.527167 +time/saving (s) 0.00347986 +time/training (s) 7.699 +time/epoch (s) 10.8376 +time/total (s) 6725.42 +Epoch -353 +---------------------------------- --------------- +2022-05-10 15:03:02.036602 PDT | [1] Epoch -352 finished +---------------------------------- --------------- +epoch -352 +replay_buffer/size 999033 +trainer/num train calls 649000 +trainer/Policy Loss -19.6412 +trainer/Log Pis Mean 25.2508 +trainer/Log Pis Std 13.5659 +trainer/Log Pis Max 69.5098 +trainer/Log Pis Min -5.18166 +trainer/policy/mean Mean -0.0199391 +trainer/policy/mean Std 0.910755 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8277 +trainer/policy/normal/std Std 0.660085 +trainer/policy/normal/std Max 5.74976 +trainer/policy/normal/std Min 0.329677 +trainer/policy/normal/log_std Mean 1.00037 +trainer/policy/normal/log_std Std 0.315415 +trainer/policy/normal/log_std Max 1.74916 +trainer/policy/normal/log_std Min -1.10964 +eval/num steps total 647133 +eval/num paths total 649 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.146548 +eval/Actions Std 0.865216 +eval/Actions Max 0.999993 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6201 +time/logging (s) 0.00373952 +time/sampling batch (s) 0.778012 +time/saving (s) 0.00349867 +time/training (s) 7.72849 +time/epoch (s) 11.1338 +time/total (s) 6736.56 +Epoch -352 +---------------------------------- --------------- +2022-05-10 15:03:12.385290 PDT | [1] Epoch -351 finished +---------------------------------- --------------- +epoch -351 +replay_buffer/size 999033 +trainer/num train calls 650000 +trainer/Policy Loss -19.4086 +trainer/Log Pis Mean 24.6841 +trainer/Log Pis Std 13.2854 +trainer/Log Pis Max 73.7022 +trainer/Log Pis Min -4.68845 +trainer/policy/mean Mean -0.0524397 +trainer/policy/mean Std 0.906741 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81894 +trainer/policy/normal/std Std 0.651979 +trainer/policy/normal/std Max 5.89845 +trainer/policy/normal/std Min 0.343709 +trainer/policy/normal/log_std Mean 0.997737 +trainer/policy/normal/log_std Std 0.313463 +trainer/policy/normal/log_std Max 1.77469 +trainer/policy/normal/log_std Min -1.06796 +eval/num steps total 648133 +eval/num paths total 650 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.205245 +eval/Actions Std 0.842925 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52785 +time/logging (s) 0.00404922 +time/sampling batch (s) 0.277968 +time/saving (s) 0.00375152 +time/training (s) 7.51193 +time/epoch (s) 10.3255 +time/total (s) 6746.89 +Epoch -351 +---------------------------------- --------------- +2022-05-10 15:03:23.798775 PDT | [1] Epoch -350 finished +---------------------------------- --------------- +epoch -350 +replay_buffer/size 999033 +trainer/num train calls 651000 +trainer/Policy Loss -20.068 +trainer/Log Pis Mean 24.1723 +trainer/Log Pis Std 13.2648 +trainer/Log Pis Max 72.9765 +trainer/Log Pis Min -7.49828 +trainer/policy/mean Mean -0.0282787 +trainer/policy/mean Std 0.897973 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85775 +trainer/policy/normal/std Std 0.67612 +trainer/policy/normal/std Max 6.83499 +trainer/policy/normal/std Min 0.245389 +trainer/policy/normal/log_std Mean 1.00988 +trainer/policy/normal/log_std Std 0.320128 +trainer/policy/normal/log_std Max 1.92206 +trainer/policy/normal/log_std Min -1.40491 +eval/num steps total 649133 +eval/num paths total 651 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232758 +eval/Actions Std 0.906677 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58831 +time/logging (s) 0.00377148 +time/sampling batch (s) 0.7776 +time/saving (s) 0.00365066 +time/training (s) 8.01621 +time/epoch (s) 11.3895 +time/total (s) 6758.28 +Epoch -350 +---------------------------------- --------------- +2022-05-10 15:03:33.798704 PDT | [1] Epoch -349 finished +---------------------------------- --------------- +epoch -349 +replay_buffer/size 999033 +trainer/num train calls 652000 +trainer/Policy Loss -20.0104 +trainer/Log Pis Mean 24.1889 +trainer/Log Pis Std 12.8909 +trainer/Log Pis Max 67.8896 +trainer/Log Pis Min -8.15355 +trainer/policy/mean Mean -0.0304656 +trainer/policy/mean Std 0.909188 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.81992 +trainer/policy/normal/std Std 0.682345 +trainer/policy/normal/std Max 6.20122 +trainer/policy/normal/std Min 0.264689 +trainer/policy/normal/log_std Mean 0.993172 +trainer/policy/normal/log_std Std 0.335421 +trainer/policy/normal/log_std Max 1.82475 +trainer/policy/normal/log_std Min -1.3292 +eval/num steps total 650133 +eval/num paths total 652 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0446934 +eval/Actions Std 0.91516 +eval/Actions Max 0.999991 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52664 +time/logging (s) 0.00382948 +time/sampling batch (s) 0.527374 +time/saving (s) 0.00340374 +time/training (s) 6.91536 +time/epoch (s) 9.97661 +time/total (s) 6768.26 +Epoch -349 +---------------------------------- --------------- +2022-05-10 15:03:44.791432 PDT | [1] Epoch -348 finished +---------------------------------- --------------- +epoch -348 +replay_buffer/size 999033 +trainer/num train calls 653000 +trainer/Policy Loss -20.4045 +trainer/Log Pis Mean 24.5748 +trainer/Log Pis Std 13.6444 +trainer/Log Pis Max 74.1626 +trainer/Log Pis Min -11.4255 +trainer/policy/mean Mean -0.0353708 +trainer/policy/mean Std 0.906902 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.80385 +trainer/policy/normal/std Std 0.651094 +trainer/policy/normal/std Max 5.59041 +trainer/policy/normal/std Min 0.295982 +trainer/policy/normal/log_std Mean 0.992222 +trainer/policy/normal/log_std Std 0.314097 +trainer/policy/normal/log_std Max 1.72105 +trainer/policy/normal/log_std Min -1.21746 +eval/num steps total 651133 +eval/num paths total 653 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.108389 +eval/Actions Std 0.894317 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67537 +time/logging (s) 0.00367003 +time/sampling batch (s) 0.276281 +time/saving (s) 0.00337855 +time/training (s) 8.01066 +time/epoch (s) 10.9694 +time/total (s) 6779.23 +Epoch -348 +---------------------------------- --------------- +2022-05-10 15:03:55.086320 PDT | [1] Epoch -347 finished +---------------------------------- --------------- +epoch -347 +replay_buffer/size 999033 +trainer/num train calls 654000 +trainer/Policy Loss -19.1742 +trainer/Log Pis Mean 24.488 +trainer/Log Pis Std 12.6861 +trainer/Log Pis Max 66.8094 +trainer/Log Pis Min -7.10218 +trainer/policy/mean Mean -0.0373961 +trainer/policy/mean Std 0.908323 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79706 +trainer/policy/normal/std Std 0.626616 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.362932 +trainer/policy/normal/log_std Mean 0.994314 +trainer/policy/normal/log_std Std 0.290804 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.01354 +eval/num steps total 652133 +eval/num paths total 654 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0224739 +eval/Actions Std 0.911148 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61074 +time/logging (s) 0.00375329 +time/sampling batch (s) 0.317955 +time/saving (s) 0.00364785 +time/training (s) 7.33503 +time/epoch (s) 10.2711 +time/total (s) 6789.51 +Epoch -347 +---------------------------------- --------------- +2022-05-10 15:04:06.602027 PDT | [1] Epoch -346 finished +---------------------------------- --------------- +epoch -346 +replay_buffer/size 999033 +trainer/num train calls 655000 +trainer/Policy Loss -18.7756 +trainer/Log Pis Mean 23.5669 +trainer/Log Pis Std 13.5827 +trainer/Log Pis Max 69.2442 +trainer/Log Pis Min -8.62705 +trainer/policy/mean Mean -0.0305066 +trainer/policy/mean Std 0.905618 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80603 +trainer/policy/normal/std Std 0.654809 +trainer/policy/normal/std Max 6.33035 +trainer/policy/normal/std Min 0.236999 +trainer/policy/normal/log_std Mean 0.9914 +trainer/policy/normal/log_std Std 0.323924 +trainer/policy/normal/log_std Max 1.84536 +trainer/policy/normal/log_std Min -1.4397 +eval/num steps total 653133 +eval/num paths total 655 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0856457 +eval/Actions Std 0.910943 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87235 +time/logging (s) 0.0040929 +time/sampling batch (s) 1.07629 +time/saving (s) 0.00382824 +time/training (s) 7.53522 +time/epoch (s) 11.4918 +time/total (s) 6801 +Epoch -346 +---------------------------------- --------------- +2022-05-10 15:04:16.493155 PDT | [1] Epoch -345 finished +---------------------------------- --------------- +epoch -345 +replay_buffer/size 999033 +trainer/num train calls 656000 +trainer/Policy Loss -18.4848 +trainer/Log Pis Mean 24.7396 +trainer/Log Pis Std 13.3808 +trainer/Log Pis Max 66.2944 +trainer/Log Pis Min -5.81385 +trainer/policy/mean Mean -0.0308122 +trainer/policy/mean Std 0.907692 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87126 +trainer/policy/normal/std Std 0.697046 +trainer/policy/normal/std Max 7.33757 +trainer/policy/normal/std Min 0.187942 +trainer/policy/normal/log_std Mean 1.00893 +trainer/policy/normal/log_std Std 0.351545 +trainer/policy/normal/log_std Max 1.99301 +trainer/policy/normal/log_std Min -1.67162 +eval/num steps total 654133 +eval/num paths total 656 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0505347 +eval/Actions Std 0.906319 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7595 +time/logging (s) 0.00382482 +time/sampling batch (s) 0.276394 +time/saving (s) 0.00377791 +time/training (s) 6.82339 +time/epoch (s) 9.86689 +time/total (s) 6810.87 +Epoch -345 +---------------------------------- --------------- +2022-05-10 15:04:26.071156 PDT | [1] Epoch -344 finished +---------------------------------- --------------- +epoch -344 +replay_buffer/size 999033 +trainer/num train calls 657000 +trainer/Policy Loss -18.7811 +trainer/Log Pis Mean 24.0208 +trainer/Log Pis Std 12.7609 +trainer/Log Pis Max 67.2507 +trainer/Log Pis Min -10.6224 +trainer/policy/mean Mean -0.00896119 +trainer/policy/mean Std 0.903432 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.79021 +trainer/policy/normal/std Std 0.674791 +trainer/policy/normal/std Max 6.90577 +trainer/policy/normal/std Min 0.243112 +trainer/policy/normal/log_std Mean 0.982904 +trainer/policy/normal/log_std Std 0.334245 +trainer/policy/normal/log_std Max 1.93236 +trainer/policy/normal/log_std Min -1.41423 +eval/num steps total 655133 +eval/num paths total 657 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0164231 +eval/Actions Std 0.93083 +eval/Actions Max 0.999982 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53643 +time/logging (s) 0.00382603 +time/sampling batch (s) 0.276037 +time/saving (s) 0.00355395 +time/training (s) 6.73466 +time/epoch (s) 9.5545 +time/total (s) 6820.43 +Epoch -344 +---------------------------------- --------------- +2022-05-10 15:04:36.067680 PDT | [1] Epoch -343 finished +---------------------------------- --------------- +epoch -343 +replay_buffer/size 999033 +trainer/num train calls 658000 +trainer/Policy Loss -19.0045 +trainer/Log Pis Mean 24.61 +trainer/Log Pis Std 13.5519 +trainer/Log Pis Max 69.4988 +trainer/Log Pis Min -5.82592 +trainer/policy/mean Mean -0.0324197 +trainer/policy/mean Std 0.90658 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82658 +trainer/policy/normal/std Std 0.659316 +trainer/policy/normal/std Max 5.75175 +trainer/policy/normal/std Min 0.302484 +trainer/policy/normal/log_std Mean 0.998824 +trainer/policy/normal/log_std Std 0.323 +trainer/policy/normal/log_std Max 1.7495 +trainer/policy/normal/log_std Min -1.19573 +eval/num steps total 656133 +eval/num paths total 658 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243424 +eval/Actions Std 0.877655 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68702 +time/logging (s) 0.00366335 +time/sampling batch (s) 0.776932 +time/saving (s) 0.00343171 +time/training (s) 6.50192 +time/epoch (s) 9.97297 +time/total (s) 6830.41 +Epoch -343 +---------------------------------- --------------- +2022-05-10 15:04:46.037304 PDT | [1] Epoch -342 finished +---------------------------------- --------------- +epoch -342 +replay_buffer/size 999033 +trainer/num train calls 659000 +trainer/Policy Loss -20.2825 +trainer/Log Pis Mean 24.7485 +trainer/Log Pis Std 13.3673 +trainer/Log Pis Max 69.5796 +trainer/Log Pis Min -7.57797 +trainer/policy/mean Mean -0.0135347 +trainer/policy/mean Std 0.902606 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77586 +trainer/policy/normal/std Std 0.676044 +trainer/policy/normal/std Max 7.07265 +trainer/policy/normal/std Min 0.249752 +trainer/policy/normal/log_std Mean 0.978762 +trainer/policy/normal/log_std Std 0.327921 +trainer/policy/normal/log_std Max 1.95623 +trainer/policy/normal/log_std Min -1.38729 +eval/num steps total 656715 +eval/num paths total 659 +eval/path length Mean 582 +eval/path length Std 0 +eval/path length Max 582 +eval/path length Min 582 +eval/Rewards Mean 0.00171821 +eval/Rewards Std 0.0414157 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.00650831 +eval/Actions Std 0.901265 +eval/Actions Max 0.999997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.29568 +time/logging (s) 0.00264462 +time/sampling batch (s) 0.280606 +time/saving (s) 0.00351089 +time/training (s) 7.36277 +time/epoch (s) 9.94521 +time/total (s) 6840.36 +Epoch -342 +---------------------------------- --------------- +2022-05-10 15:04:55.492950 PDT | [1] Epoch -341 finished +---------------------------------- --------------- +epoch -341 +replay_buffer/size 999033 +trainer/num train calls 660000 +trainer/Policy Loss -19.869 +trainer/Log Pis Mean 24.0651 +trainer/Log Pis Std 13.9451 +trainer/Log Pis Max 77.0844 +trainer/Log Pis Min -5.66131 +trainer/policy/mean Mean -0.0413146 +trainer/policy/mean Std 0.906004 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80889 +trainer/policy/normal/std Std 0.698225 +trainer/policy/normal/std Max 6.03016 +trainer/policy/normal/std Min 0.275753 +trainer/policy/normal/log_std Mean 0.986185 +trainer/policy/normal/log_std Std 0.349416 +trainer/policy/normal/log_std Max 1.79677 +trainer/policy/normal/log_std Min -1.28825 +eval/num steps total 657715 +eval/num paths total 660 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.537829 +eval/Actions Std 0.764578 +eval/Actions Max 0.999991 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73736 +time/logging (s) 0.0037058 +time/sampling batch (s) 0.526679 +time/saving (s) 0.00340809 +time/training (s) 6.16235 +time/epoch (s) 9.4335 +time/total (s) 6849.79 +Epoch -341 +---------------------------------- --------------- +2022-05-10 15:05:05.636618 PDT | [1] Epoch -340 finished +---------------------------------- --------------- +epoch -340 +replay_buffer/size 999033 +trainer/num train calls 661000 +trainer/Policy Loss -19.2758 +trainer/Log Pis Mean 25.3283 +trainer/Log Pis Std 13.244 +trainer/Log Pis Max 69.7775 +trainer/Log Pis Min -7.46164 +trainer/policy/mean Mean -0.0450986 +trainer/policy/mean Std 0.909172 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81089 +trainer/policy/normal/std Std 0.665082 +trainer/policy/normal/std Max 5.5589 +trainer/policy/normal/std Min 0.25811 +trainer/policy/normal/log_std Mean 0.992836 +trainer/policy/normal/log_std Std 0.322812 +trainer/policy/normal/log_std Max 1.7154 +trainer/policy/normal/log_std Min -1.35437 +eval/num steps total 658715 +eval/num paths total 661 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.136928 +eval/Actions Std 0.897277 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66355 +time/logging (s) 0.00418945 +time/sampling batch (s) 0.778994 +time/saving (s) 0.0039584 +time/training (s) 6.66994 +time/epoch (s) 10.1206 +time/total (s) 6859.92 +Epoch -340 +---------------------------------- --------------- +2022-05-10 15:05:16.017532 PDT | [1] Epoch -339 finished +---------------------------------- --------------- +epoch -339 +replay_buffer/size 999033 +trainer/num train calls 662000 +trainer/Policy Loss -19.3092 +trainer/Log Pis Mean 24.4 +trainer/Log Pis Std 13.3757 +trainer/Log Pis Max 86.4564 +trainer/Log Pis Min -14.7232 +trainer/policy/mean Mean -0.0276585 +trainer/policy/mean Std 0.908525 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8302 +trainer/policy/normal/std Std 0.650473 +trainer/policy/normal/std Max 5.16114 +trainer/policy/normal/std Min 0.276252 +trainer/policy/normal/log_std Mean 1.00144 +trainer/policy/normal/log_std Std 0.317498 +trainer/policy/normal/log_std Max 1.64116 +trainer/policy/normal/log_std Min -1.28644 +eval/num steps total 659715 +eval/num paths total 662 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0912727 +eval/Actions Std 0.883065 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50954 +time/logging (s) 0.00395813 +time/sampling batch (s) 0.282639 +time/saving (s) 0.00359272 +time/training (s) 7.55688 +time/epoch (s) 10.3566 +time/total (s) 6870.28 +Epoch -339 +---------------------------------- --------------- +2022-05-10 15:05:27.275256 PDT | [1] Epoch -338 finished +---------------------------------- --------------- +epoch -338 +replay_buffer/size 999033 +trainer/num train calls 663000 +trainer/Policy Loss -19.9827 +trainer/Log Pis Mean 24.4248 +trainer/Log Pis Std 13.1125 +trainer/Log Pis Max 67.7471 +trainer/Log Pis Min -9.90776 +trainer/policy/mean Mean -0.0430236 +trainer/policy/mean Std 0.907883 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80671 +trainer/policy/normal/std Std 0.650661 +trainer/policy/normal/std Max 6.03521 +trainer/policy/normal/std Min 0.304202 +trainer/policy/normal/log_std Mean 0.993679 +trainer/policy/normal/log_std Std 0.31161 +trainer/policy/normal/log_std Max 1.79761 +trainer/policy/normal/log_std Min -1.19006 +eval/num steps total 660715 +eval/num paths total 663 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.329846 +eval/Actions Std 0.84641 +eval/Actions Max 0.999969 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65301 +time/logging (s) 0.00381359 +time/sampling batch (s) 0.28303 +time/saving (s) 0.00348536 +time/training (s) 8.29032 +time/epoch (s) 11.2337 +time/total (s) 6881.51 +Epoch -338 +---------------------------------- --------------- +2022-05-10 15:05:36.937944 PDT | [1] Epoch -337 finished +---------------------------------- ---------------- +epoch -337 +replay_buffer/size 999033 +trainer/num train calls 664000 +trainer/Policy Loss -19.4718 +trainer/Log Pis Mean 24.915 +trainer/Log Pis Std 12.7343 +trainer/Log Pis Max 74.5282 +trainer/Log Pis Min -4.69343 +trainer/policy/mean Mean -0.000628963 +trainer/policy/mean Std 0.908858 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79983 +trainer/policy/normal/std Std 0.651655 +trainer/policy/normal/std Max 7.10867 +trainer/policy/normal/std Min 0.33151 +trainer/policy/normal/log_std Mean 0.990798 +trainer/policy/normal/log_std Std 0.313642 +trainer/policy/normal/log_std Max 1.96131 +trainer/policy/normal/log_std Min -1.1041 +eval/num steps total 661715 +eval/num paths total 664 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.275842 +eval/Actions Std 0.908576 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52623 +time/logging (s) 0.0037916 +time/sampling batch (s) 0.284131 +time/saving (s) 0.00346358 +time/training (s) 6.82125 +time/epoch (s) 9.63887 +time/total (s) 6891.15 +Epoch -337 +---------------------------------- ---------------- +2022-05-10 15:05:48.243813 PDT | [1] Epoch -336 finished +---------------------------------- --------------- +epoch -336 +replay_buffer/size 999033 +trainer/num train calls 665000 +trainer/Policy Loss -19.3159 +trainer/Log Pis Mean 24.2908 +trainer/Log Pis Std 12.6904 +trainer/Log Pis Max 75.2485 +trainer/Log Pis Min -13.6537 +trainer/policy/mean Mean -0.0153009 +trainer/policy/mean Std 0.905076 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78829 +trainer/policy/normal/std Std 0.661703 +trainer/policy/normal/std Max 4.9543 +trainer/policy/normal/std Min 0.293879 +trainer/policy/normal/log_std Mean 0.981932 +trainer/policy/normal/log_std Std 0.339114 +trainer/policy/normal/log_std Max 1.60026 +trainer/policy/normal/log_std Min -1.22459 +eval/num steps total 662715 +eval/num paths total 665 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.452529 +eval/Actions Std 0.80796 +eval/Actions Max 0.999993 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60964 +time/logging (s) 0.00374871 +time/sampling batch (s) 0.532332 +time/saving (s) 0.0034405 +time/training (s) 8.13303 +time/epoch (s) 11.2822 +time/total (s) 6902.44 +Epoch -336 +---------------------------------- --------------- +2022-05-10 15:05:58.150128 PDT | [1] Epoch -335 finished +---------------------------------- --------------- +epoch -335 +replay_buffer/size 999033 +trainer/num train calls 666000 +trainer/Policy Loss -20.1708 +trainer/Log Pis Mean 25.2355 +trainer/Log Pis Std 13.5186 +trainer/Log Pis Max 67.5718 +trainer/Log Pis Min -8.25057 +trainer/policy/mean Mean -0.0392584 +trainer/policy/mean Std 0.906379 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83875 +trainer/policy/normal/std Std 0.66614 +trainer/policy/normal/std Max 5.63702 +trainer/policy/normal/std Min 0.318967 +trainer/policy/normal/log_std Mean 1.00453 +trainer/policy/normal/log_std Std 0.312415 +trainer/policy/normal/log_std Max 1.72936 +trainer/policy/normal/log_std Min -1.14267 +eval/num steps total 663715 +eval/num paths total 666 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0252528 +eval/Actions Std 0.908362 +eval/Actions Max 0.999977 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61273 +time/logging (s) 0.00408075 +time/sampling batch (s) 0.278968 +time/saving (s) 0.00384896 +time/training (s) 6.98348 +time/epoch (s) 9.88312 +time/total (s) 6912.32 +Epoch -335 +---------------------------------- --------------- +2022-05-10 15:06:08.140509 PDT | [1] Epoch -334 finished +---------------------------------- --------------- +epoch -334 +replay_buffer/size 999033 +trainer/num train calls 667000 +trainer/Policy Loss -19.7905 +trainer/Log Pis Mean 24.7222 +trainer/Log Pis Std 13.5885 +trainer/Log Pis Max 72.8491 +trainer/Log Pis Min -9.3603 +trainer/policy/mean Mean -0.0453004 +trainer/policy/mean Std 0.903387 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78004 +trainer/policy/normal/std Std 0.653288 +trainer/policy/normal/std Max 6.45866 +trainer/policy/normal/std Min 0.313473 +trainer/policy/normal/log_std Mean 0.983955 +trainer/policy/normal/log_std Std 0.309609 +trainer/policy/normal/log_std Max 1.86542 +trainer/policy/normal/log_std Min -1.16004 +eval/num steps total 664715 +eval/num paths total 667 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00352723 +eval/Actions Std 0.906203 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55236 +time/logging (s) 0.00381003 +time/sampling batch (s) 0.530178 +time/saving (s) 0.00375855 +time/training (s) 6.87601 +time/epoch (s) 9.96611 +time/total (s) 6922.29 +Epoch -334 +---------------------------------- --------------- +2022-05-10 15:06:19.516999 PDT | [1] Epoch -333 finished +---------------------------------- --------------- +epoch -333 +replay_buffer/size 999033 +trainer/num train calls 668000 +trainer/Policy Loss -19.3223 +trainer/Log Pis Mean 24.1034 +trainer/Log Pis Std 14.0429 +trainer/Log Pis Max 86.1509 +trainer/Log Pis Min -11.5395 +trainer/policy/mean Mean -0.0480332 +trainer/policy/mean Std 0.902859 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86049 +trainer/policy/normal/std Std 0.652102 +trainer/policy/normal/std Max 5.897 +trainer/policy/normal/std Min 0.291667 +trainer/policy/normal/log_std Mean 1.01442 +trainer/policy/normal/log_std Std 0.303225 +trainer/policy/normal/log_std Max 1.77444 +trainer/policy/normal/log_std Min -1.23214 +eval/num steps total 665715 +eval/num paths total 668 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.092654 +eval/Actions Std 0.916285 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63984 +time/logging (s) 0.00365929 +time/sampling batch (s) 0.27975 +time/saving (s) 0.0034577 +time/training (s) 8.42593 +time/epoch (s) 11.3526 +time/total (s) 6933.65 +Epoch -333 +---------------------------------- --------------- +2022-05-10 15:06:30.660221 PDT | [1] Epoch -332 finished +---------------------------------- --------------- +epoch -332 +replay_buffer/size 999033 +trainer/num train calls 669000 +trainer/Policy Loss -18.8044 +trainer/Log Pis Mean 24.7316 +trainer/Log Pis Std 13.69 +trainer/Log Pis Max 70.439 +trainer/Log Pis Min -5.97732 +trainer/policy/mean Mean -0.0454821 +trainer/policy/mean Std 0.904765 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.85264 +trainer/policy/normal/std Std 0.671573 +trainer/policy/normal/std Max 6.62244 +trainer/policy/normal/std Min 0.326614 +trainer/policy/normal/log_std Mean 1.00893 +trainer/policy/normal/log_std Std 0.315596 +trainer/policy/normal/log_std Max 1.89046 +trainer/policy/normal/log_std Min -1.11898 +eval/num steps total 666715 +eval/num paths total 669 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0356795 +eval/Actions Std 0.903726 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57252 +time/logging (s) 0.00379386 +time/sampling batch (s) 0.527773 +time/saving (s) 0.00344621 +time/training (s) 8.01259 +time/epoch (s) 11.1201 +time/total (s) 6944.77 +Epoch -332 +---------------------------------- --------------- +2022-05-10 15:06:41.253206 PDT | [1] Epoch -331 finished +---------------------------------- --------------- +epoch -331 +replay_buffer/size 999033 +trainer/num train calls 670000 +trainer/Policy Loss -18.9813 +trainer/Log Pis Mean 24.8482 +trainer/Log Pis Std 13.4828 +trainer/Log Pis Max 64.6266 +trainer/Log Pis Min -13.5467 +trainer/policy/mean Mean -0.0381761 +trainer/policy/mean Std 0.907585 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.807 +trainer/policy/normal/std Std 0.676272 +trainer/policy/normal/std Max 5.38513 +trainer/policy/normal/std Min 0.267991 +trainer/policy/normal/log_std Mean 0.987457 +trainer/policy/normal/log_std Std 0.34296 +trainer/policy/normal/log_std Max 1.68364 +trainer/policy/normal/log_std Min -1.3168 +eval/num steps total 667715 +eval/num paths total 670 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.159854 +eval/Actions Std 0.840341 +eval/Actions Max 0.999982 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65681 +time/logging (s) 0.00374725 +time/sampling batch (s) 0.525493 +time/saving (s) 0.00344988 +time/training (s) 7.38016 +time/epoch (s) 10.5697 +time/total (s) 6955.35 +Epoch -331 +---------------------------------- --------------- +2022-05-10 15:06:51.445168 PDT | [1] Epoch -330 finished +---------------------------------- --------------- +epoch -330 +replay_buffer/size 999033 +trainer/num train calls 671000 +trainer/Policy Loss -20.2834 +trainer/Log Pis Mean 24.2082 +trainer/Log Pis Std 13.806 +trainer/Log Pis Max 67.3992 +trainer/Log Pis Min -9.14622 +trainer/policy/mean Mean -0.0385618 +trainer/policy/mean Std 0.90708 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84786 +trainer/policy/normal/std Std 0.654734 +trainer/policy/normal/std Max 5.24632 +trainer/policy/normal/std Min 0.216687 +trainer/policy/normal/log_std Mean 1.0086 +trainer/policy/normal/log_std Std 0.310634 +trainer/policy/normal/log_std Max 1.65753 +trainer/policy/normal/log_std Min -1.5293 +eval/num steps total 668715 +eval/num paths total 671 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.351773 +eval/Actions Std 0.859465 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6964 +time/logging (s) 0.00367921 +time/sampling batch (s) 0.776232 +time/saving (s) 0.00343301 +time/training (s) 6.6888 +time/epoch (s) 10.1685 +time/total (s) 6965.52 +Epoch -330 +---------------------------------- --------------- +2022-05-10 15:07:01.894805 PDT | [1] Epoch -329 finished +---------------------------------- --------------- +epoch -329 +replay_buffer/size 999033 +trainer/num train calls 672000 +trainer/Policy Loss -19.447 +trainer/Log Pis Mean 25.9081 +trainer/Log Pis Std 13.2667 +trainer/Log Pis Max 80.9782 +trainer/Log Pis Min -5.37782 +trainer/policy/mean Mean -0.0255551 +trainer/policy/mean Std 0.907489 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81647 +trainer/policy/normal/std Std 0.692115 +trainer/policy/normal/std Max 5.47067 +trainer/policy/normal/std Min 0.24845 +trainer/policy/normal/log_std Mean 0.988654 +trainer/policy/normal/log_std Std 0.3531 +trainer/policy/normal/log_std Max 1.6994 +trainer/policy/normal/log_std Min -1.39251 +eval/num steps total 669715 +eval/num paths total 672 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0700884 +eval/Actions Std 0.904586 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66403 +time/logging (s) 0.00431761 +time/sampling batch (s) 0.280591 +time/saving (s) 0.00397077 +time/training (s) 7.4736 +time/epoch (s) 10.4265 +time/total (s) 6975.95 +Epoch -329 +---------------------------------- --------------- +2022-05-10 15:07:12.029952 PDT | [1] Epoch -328 finished +---------------------------------- --------------- +epoch -328 +replay_buffer/size 999033 +trainer/num train calls 673000 +trainer/Policy Loss -19.4502 +trainer/Log Pis Mean 24.2764 +trainer/Log Pis Std 13.1117 +trainer/Log Pis Max 69.906 +trainer/Log Pis Min -8.01968 +trainer/policy/mean Mean -0.023339 +trainer/policy/mean Std 0.910079 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77853 +trainer/policy/normal/std Std 0.640426 +trainer/policy/normal/std Max 6.32623 +trainer/policy/normal/std Min 0.258319 +trainer/policy/normal/log_std Mean 0.983777 +trainer/policy/normal/log_std Std 0.312409 +trainer/policy/normal/log_std Max 1.8447 +trainer/policy/normal/log_std Min -1.35356 +eval/num steps total 670715 +eval/num paths total 673 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.220068 +eval/Actions Std 0.930083 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7519 +time/logging (s) 0.00378441 +time/sampling batch (s) 0.281302 +time/saving (s) 0.00367826 +time/training (s) 7.06976 +time/epoch (s) 10.1104 +time/total (s) 6986.06 +Epoch -328 +---------------------------------- --------------- +2022-05-10 15:07:22.802503 PDT | [1] Epoch -327 finished +---------------------------------- --------------- +epoch -327 +replay_buffer/size 999033 +trainer/num train calls 674000 +trainer/Policy Loss -21.0522 +trainer/Log Pis Mean 24.806 +trainer/Log Pis Std 12.6514 +trainer/Log Pis Max 69.2467 +trainer/Log Pis Min -9.54756 +trainer/policy/mean Mean -0.0342776 +trainer/policy/mean Std 0.909772 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.77571 +trainer/policy/normal/std Std 0.668341 +trainer/policy/normal/std Max 5.68365 +trainer/policy/normal/std Min 0.256734 +trainer/policy/normal/log_std Mean 0.976529 +trainer/policy/normal/log_std Std 0.343082 +trainer/policy/normal/log_std Max 1.73759 +trainer/policy/normal/log_std Min -1.35971 +eval/num steps total 671715 +eval/num paths total 674 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0887845 +eval/Actions Std 0.917349 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63844 +time/logging (s) 0.00366362 +time/sampling batch (s) 0.280253 +time/saving (s) 0.00335557 +time/training (s) 7.823 +time/epoch (s) 10.7487 +time/total (s) 6996.81 +Epoch -327 +---------------------------------- --------------- +2022-05-10 15:07:33.135070 PDT | [1] Epoch -326 finished +---------------------------------- --------------- +epoch -326 +replay_buffer/size 999033 +trainer/num train calls 675000 +trainer/Policy Loss -19.2844 +trainer/Log Pis Mean 23.6997 +trainer/Log Pis Std 12.904 +trainer/Log Pis Max 64.5277 +trainer/Log Pis Min -9.69572 +trainer/policy/mean Mean -0.0251254 +trainer/policy/mean Std 0.905429 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83337 +trainer/policy/normal/std Std 0.653882 +trainer/policy/normal/std Max 5.7502 +trainer/policy/normal/std Min 0.304393 +trainer/policy/normal/log_std Mean 1.0034 +trainer/policy/normal/log_std Std 0.310488 +trainer/policy/normal/log_std Max 1.74923 +trainer/policy/normal/log_std Min -1.18944 +eval/num steps total 672715 +eval/num paths total 675 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.027012 +eval/Actions Std 0.912736 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60989 +time/logging (s) 0.00366486 +time/sampling batch (s) 0.277264 +time/saving (s) 0.00338472 +time/training (s) 7.41491 +time/epoch (s) 10.3091 +time/total (s) 7007.12 +Epoch -326 +---------------------------------- --------------- +2022-05-10 15:07:43.474257 PDT | [1] Epoch -325 finished +---------------------------------- --------------- +epoch -325 +replay_buffer/size 999033 +trainer/num train calls 676000 +trainer/Policy Loss -19.1997 +trainer/Log Pis Mean 24.203 +trainer/Log Pis Std 13.1682 +trainer/Log Pis Max 65.9161 +trainer/Log Pis Min -6.82731 +trainer/policy/mean Mean -0.0313895 +trainer/policy/mean Std 0.904505 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.84816 +trainer/policy/normal/std Std 0.677624 +trainer/policy/normal/std Max 5.62702 +trainer/policy/normal/std Min 0.277801 +trainer/policy/normal/log_std Mean 1.00494 +trainer/policy/normal/log_std Std 0.32877 +trainer/policy/normal/log_std Max 1.72758 +trainer/policy/normal/log_std Min -1.28085 +eval/num steps total 673715 +eval/num paths total 676 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.187373 +eval/Actions Std 0.888231 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46275 +time/logging (s) 0.00364671 +time/sampling batch (s) 0.532643 +time/saving (s) 0.00343679 +time/training (s) 7.31315 +time/epoch (s) 10.3156 +time/total (s) 7017.44 +Epoch -325 +---------------------------------- --------------- +2022-05-10 15:07:53.298762 PDT | [1] Epoch -324 finished +---------------------------------- --------------- +epoch -324 +replay_buffer/size 999033 +trainer/num train calls 677000 +trainer/Policy Loss -19.6107 +trainer/Log Pis Mean 24.1202 +trainer/Log Pis Std 13.0336 +trainer/Log Pis Max 63.4898 +trainer/Log Pis Min -10.9263 +trainer/policy/mean Mean -0.0381321 +trainer/policy/mean Std 0.910753 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81567 +trainer/policy/normal/std Std 0.639739 +trainer/policy/normal/std Max 5.94878 +trainer/policy/normal/std Min 0.274709 +trainer/policy/normal/log_std Mean 0.997795 +trainer/policy/normal/log_std Std 0.309577 +trainer/policy/normal/log_std Max 1.78319 +trainer/policy/normal/log_std Min -1.29204 +eval/num steps total 674715 +eval/num paths total 677 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.118819 +eval/Actions Std 0.918244 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54937 +time/logging (s) 0.00374035 +time/sampling batch (s) 0.278689 +time/saving (s) 0.00340742 +time/training (s) 6.96574 +time/epoch (s) 9.80095 +time/total (s) 7027.25 +Epoch -324 +---------------------------------- --------------- +2022-05-10 15:08:03.787439 PDT | [1] Epoch -323 finished +---------------------------------- --------------- +epoch -323 +replay_buffer/size 999033 +trainer/num train calls 678000 +trainer/Policy Loss -19.2118 +trainer/Log Pis Mean 23.1497 +trainer/Log Pis Std 12.65 +trainer/Log Pis Max 72.662 +trainer/Log Pis Min -5.93188 +trainer/policy/mean Mean -0.0146253 +trainer/policy/mean Std 0.902135 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75108 +trainer/policy/normal/std Std 0.684395 +trainer/policy/normal/std Max 7.17554 +trainer/policy/normal/std Min 0.242283 +trainer/policy/normal/log_std Mean 0.967415 +trainer/policy/normal/log_std Std 0.33655 +trainer/policy/normal/log_std Max 1.97068 +trainer/policy/normal/log_std Min -1.41765 +eval/num steps total 675715 +eval/num paths total 678 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.298384 +eval/Actions Std 0.853356 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62481 +time/logging (s) 0.00394648 +time/sampling batch (s) 0.279537 +time/saving (s) 0.00402353 +time/training (s) 7.55304 +time/epoch (s) 10.4654 +time/total (s) 7037.71 +Epoch -323 +---------------------------------- --------------- +2022-05-10 15:08:14.805476 PDT | [1] Epoch -322 finished +---------------------------------- --------------- +epoch -322 +replay_buffer/size 999033 +trainer/num train calls 679000 +trainer/Policy Loss -20.0878 +trainer/Log Pis Mean 23.9531 +trainer/Log Pis Std 13.1155 +trainer/Log Pis Max 67.0217 +trainer/Log Pis Min -3.91487 +trainer/policy/mean Mean -0.0333041 +trainer/policy/mean Std 0.904009 +trainer/policy/mean Max 0.999972 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77033 +trainer/policy/normal/std Std 0.657374 +trainer/policy/normal/std Max 6.21359 +trainer/policy/normal/std Min 0.260558 +trainer/policy/normal/log_std Mean 0.977759 +trainer/policy/normal/log_std Std 0.324808 +trainer/policy/normal/log_std Max 1.82674 +trainer/policy/normal/log_std Min -1.34493 +eval/num steps total 676715 +eval/num paths total 679 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.11437 +eval/Actions Std 0.91638 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67311 +time/logging (s) 0.00368088 +time/sampling batch (s) 1.02774 +time/saving (s) 0.00336426 +time/training (s) 7.28599 +time/epoch (s) 10.9939 +time/total (s) 7048.71 +Epoch -322 +---------------------------------- --------------- +2022-05-10 15:08:24.905295 PDT | [1] Epoch -321 finished +---------------------------------- --------------- +epoch -321 +replay_buffer/size 999033 +trainer/num train calls 680000 +trainer/Policy Loss -19.7524 +trainer/Log Pis Mean 25.3779 +trainer/Log Pis Std 13.4041 +trainer/Log Pis Max 83.3736 +trainer/Log Pis Min -7.40238 +trainer/policy/mean Mean -0.0477787 +trainer/policy/mean Std 0.90421 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8369 +trainer/policy/normal/std Std 0.662617 +trainer/policy/normal/std Max 5.68733 +trainer/policy/normal/std Min 0.293091 +trainer/policy/normal/log_std Mean 1.00282 +trainer/policy/normal/log_std Std 0.320769 +trainer/policy/normal/log_std Max 1.73824 +trainer/policy/normal/log_std Min -1.22727 +eval/num steps total 677715 +eval/num paths total 680 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109682 +eval/Actions Std 0.907432 +eval/Actions Max 0.999997 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57389 +time/logging (s) 0.00369172 +time/sampling batch (s) 0.52811 +time/saving (s) 0.00344146 +time/training (s) 6.96721 +time/epoch (s) 10.0763 +time/total (s) 7058.79 +Epoch -321 +---------------------------------- --------------- +2022-05-10 15:08:34.854104 PDT | [1] Epoch -320 finished +---------------------------------- --------------- +epoch -320 +replay_buffer/size 999033 +trainer/num train calls 681000 +trainer/Policy Loss -18.9282 +trainer/Log Pis Mean 24.202 +trainer/Log Pis Std 12.8482 +trainer/Log Pis Max 62.0234 +trainer/Log Pis Min -8.54572 +trainer/policy/mean Mean -0.0521944 +trainer/policy/mean Std 0.903963 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.7516 +trainer/policy/normal/std Std 0.632725 +trainer/policy/normal/std Max 5.16817 +trainer/policy/normal/std Min 0.316926 +trainer/policy/normal/log_std Mean 0.973804 +trainer/policy/normal/log_std Std 0.3132 +trainer/policy/normal/log_std Max 1.64252 +trainer/policy/normal/log_std Min -1.14909 +eval/num steps total 678715 +eval/num paths total 681 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0871139 +eval/Actions Std 0.904272 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77169 +time/logging (s) 0.00381912 +time/sampling batch (s) 0.524936 +time/saving (s) 0.00343286 +time/training (s) 6.62163 +time/epoch (s) 9.92551 +time/total (s) 7068.72 +Epoch -320 +---------------------------------- --------------- +2022-05-10 15:08:44.323360 PDT | [1] Epoch -319 finished +---------------------------------- --------------- +epoch -319 +replay_buffer/size 999033 +trainer/num train calls 682000 +trainer/Policy Loss -18.6777 +trainer/Log Pis Mean 24.1703 +trainer/Log Pis Std 12.761 +trainer/Log Pis Max 72.0024 +trainer/Log Pis Min -8.62404 +trainer/policy/mean Mean -0.0487598 +trainer/policy/mean Std 0.906921 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77933 +trainer/policy/normal/std Std 0.652184 +trainer/policy/normal/std Max 5.74674 +trainer/policy/normal/std Min 0.324851 +trainer/policy/normal/log_std Mean 0.982879 +trainer/policy/normal/log_std Std 0.315821 +trainer/policy/normal/log_std Max 1.74863 +trainer/policy/normal/log_std Min -1.12439 +eval/num steps total 679715 +eval/num paths total 682 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.302903 +eval/Actions Std 0.815504 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43884 +time/logging (s) 0.00377712 +time/sampling batch (s) 0.27819 +time/saving (s) 0.00344993 +time/training (s) 6.72149 +time/epoch (s) 9.44574 +time/total (s) 7078.17 +Epoch -319 +---------------------------------- --------------- +2022-05-10 15:08:54.545313 PDT | [1] Epoch -318 finished +---------------------------------- -------------- +epoch -318 +replay_buffer/size 999033 +trainer/num train calls 683000 +trainer/Policy Loss -19.8224 +trainer/Log Pis Mean 24.389 +trainer/Log Pis Std 13.5362 +trainer/Log Pis Max 68.2743 +trainer/Log Pis Min -12.9682 +trainer/policy/mean Mean -0.0204499 +trainer/policy/mean Std 0.906705 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83801 +trainer/policy/normal/std Std 0.665828 +trainer/policy/normal/std Max 5.73536 +trainer/policy/normal/std Min 0.2712 +trainer/policy/normal/log_std Mean 1.0033 +trainer/policy/normal/log_std Std 0.318467 +trainer/policy/normal/log_std Max 1.74665 +trainer/policy/normal/log_std Min -1.3049 +eval/num steps total 680715 +eval/num paths total 683 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0441945 +eval/Actions Std 0.905314 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50096 +time/logging (s) 0.0037911 +time/sampling batch (s) 0.27921 +time/saving (s) 0.0034398 +time/training (s) 7.41079 +time/epoch (s) 10.1982 +time/total (s) 7088.37 +Epoch -318 +---------------------------------- -------------- +2022-05-10 15:09:04.157062 PDT | [1] Epoch -317 finished +---------------------------------- --------------- +epoch -317 +replay_buffer/size 999033 +trainer/num train calls 684000 +trainer/Policy Loss -19.6127 +trainer/Log Pis Mean 24.4282 +trainer/Log Pis Std 13.2086 +trainer/Log Pis Max 72.5796 +trainer/Log Pis Min -11.2156 +trainer/policy/mean Mean -0.0250645 +trainer/policy/mean Std 0.907555 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.841 +trainer/policy/normal/std Std 0.677256 +trainer/policy/normal/std Max 5.90401 +trainer/policy/normal/std Min 0.228611 +trainer/policy/normal/log_std Mean 1.00206 +trainer/policy/normal/log_std Std 0.330867 +trainer/policy/normal/log_std Max 1.77563 +trainer/policy/normal/log_std Min -1.47573 +eval/num steps total 681715 +eval/num paths total 684 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12551 +eval/Actions Std 0.889098 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59118 +time/logging (s) 0.00407341 +time/sampling batch (s) 0.528381 +time/saving (s) 0.00401896 +time/training (s) 6.46068 +time/epoch (s) 9.58834 +time/total (s) 7097.96 +Epoch -317 +---------------------------------- --------------- +2022-05-10 15:09:13.935749 PDT | [1] Epoch -316 finished +---------------------------------- --------------- +epoch -316 +replay_buffer/size 999033 +trainer/num train calls 685000 +trainer/Policy Loss -20.0906 +trainer/Log Pis Mean 24.5949 +trainer/Log Pis Std 13.7025 +trainer/Log Pis Max 68.2527 +trainer/Log Pis Min -6.10719 +trainer/policy/mean Mean -0.0359483 +trainer/policy/mean Std 0.907344 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75131 +trainer/policy/normal/std Std 0.65776 +trainer/policy/normal/std Max 6.44853 +trainer/policy/normal/std Min 0.310384 +trainer/policy/normal/log_std Mean 0.970462 +trainer/policy/normal/log_std Std 0.326405 +trainer/policy/normal/log_std Max 1.86385 +trainer/policy/normal/log_std Min -1.16995 +eval/num steps total 682715 +eval/num paths total 685 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.284678 +eval/Actions Std 0.914439 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60018 +time/logging (s) 0.00398307 +time/sampling batch (s) 0.286878 +time/saving (s) 0.00349171 +time/training (s) 6.85959 +time/epoch (s) 9.75412 +time/total (s) 7107.72 +Epoch -316 +---------------------------------- --------------- +2022-05-10 15:09:23.318413 PDT | [1] Epoch -315 finished +---------------------------------- --------------- +epoch -315 +replay_buffer/size 999033 +trainer/num train calls 686000 +trainer/Policy Loss -19.3518 +trainer/Log Pis Mean 24.4977 +trainer/Log Pis Std 13.0373 +trainer/Log Pis Max 69.8966 +trainer/Log Pis Min -12.6826 +trainer/policy/mean Mean -0.0351648 +trainer/policy/mean Std 0.90572 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77108 +trainer/policy/normal/std Std 0.63223 +trainer/policy/normal/std Max 4.94999 +trainer/policy/normal/std Min 0.296496 +trainer/policy/normal/log_std Mean 0.98071 +trainer/policy/normal/log_std Std 0.31571 +trainer/policy/normal/log_std Max 1.59939 +trainer/policy/normal/log_std Min -1.21572 +eval/num steps total 683715 +eval/num paths total 686 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.17661 +eval/Actions Std 0.93901 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4717 +time/logging (s) 0.00367744 +time/sampling batch (s) 0.278228 +time/saving (s) 0.00352324 +time/training (s) 6.60156 +time/epoch (s) 9.35869 +time/total (s) 7117.08 +Epoch -315 +---------------------------------- --------------- +2022-05-10 15:09:32.854990 PDT | [1] Epoch -314 finished +---------------------------------- --------------- +epoch -314 +replay_buffer/size 999033 +trainer/num train calls 687000 +trainer/Policy Loss -19.8009 +trainer/Log Pis Mean 24.9585 +trainer/Log Pis Std 13.8483 +trainer/Log Pis Max 61.8183 +trainer/Log Pis Min -9.34848 +trainer/policy/mean Mean -0.0358744 +trainer/policy/mean Std 0.908407 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82963 +trainer/policy/normal/std Std 0.671792 +trainer/policy/normal/std Max 5.83034 +trainer/policy/normal/std Min 0.313662 +trainer/policy/normal/log_std Mean 0.999755 +trainer/policy/normal/log_std Std 0.320512 +trainer/policy/normal/log_std Max 1.76307 +trainer/policy/normal/log_std Min -1.15944 +eval/num steps total 684715 +eval/num paths total 687 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.01478 +eval/Actions Std 0.900867 +eval/Actions Max 0.999987 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68745 +time/logging (s) 0.00375891 +time/sampling batch (s) 0.278398 +time/saving (s) 0.00353087 +time/training (s) 6.53998 +time/epoch (s) 9.51312 +time/total (s) 7126.6 +Epoch -314 +---------------------------------- --------------- +2022-05-10 15:09:43.265010 PDT | [1] Epoch -313 finished +---------------------------------- --------------- +epoch -313 +replay_buffer/size 999033 +trainer/num train calls 688000 +trainer/Policy Loss -19.0014 +trainer/Log Pis Mean 24.0224 +trainer/Log Pis Std 13.0881 +trainer/Log Pis Max 68.2897 +trainer/Log Pis Min -6.63861 +trainer/policy/mean Mean -0.0477217 +trainer/policy/mean Std 0.908606 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.78153 +trainer/policy/normal/std Std 0.654628 +trainer/policy/normal/std Max 5.78511 +trainer/policy/normal/std Min 0.207073 +trainer/policy/normal/log_std Mean 0.984025 +trainer/policy/normal/log_std Std 0.313426 +trainer/policy/normal/log_std Max 1.75529 +trainer/policy/normal/log_std Min -1.57468 +eval/num steps total 685715 +eval/num paths total 688 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106508 +eval/Actions Std 0.911437 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71707 +time/logging (s) 0.00375757 +time/sampling batch (s) 0.531323 +time/saving (s) 0.00345131 +time/training (s) 7.13055 +time/epoch (s) 10.3862 +time/total (s) 7136.98 +Epoch -313 +---------------------------------- --------------- +2022-05-10 15:09:53.838140 PDT | [1] Epoch -312 finished +---------------------------------- --------------- +epoch -312 +replay_buffer/size 999033 +trainer/num train calls 689000 +trainer/Policy Loss -19.7319 +trainer/Log Pis Mean 24.1947 +trainer/Log Pis Std 13.2677 +trainer/Log Pis Max 57.8903 +trainer/Log Pis Min -11.1372 +trainer/policy/mean Mean -0.027778 +trainer/policy/mean Std 0.908698 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.87592 +trainer/policy/normal/std Std 0.670284 +trainer/policy/normal/std Max 6.10817 +trainer/policy/normal/std Min 0.296981 +trainer/policy/normal/log_std Mean 1.01718 +trainer/policy/normal/log_std Std 0.31731 +trainer/policy/normal/log_std Max 1.80963 +trainer/policy/normal/log_std Min -1.21409 +eval/num steps total 686715 +eval/num paths total 689 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0299785 +eval/Actions Std 0.906134 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69512 +time/logging (s) 0.00424622 +time/sampling batch (s) 0.529952 +time/saving (s) 0.0039602 +time/training (s) 7.31659 +time/epoch (s) 10.5499 +time/total (s) 7147.54 +Epoch -312 +---------------------------------- --------------- +2022-05-10 15:10:04.421523 PDT | [1] Epoch -311 finished +---------------------------------- --------------- +epoch -311 +replay_buffer/size 999033 +trainer/num train calls 690000 +trainer/Policy Loss -19.2945 +trainer/Log Pis Mean 24.7397 +trainer/Log Pis Std 13.3129 +trainer/Log Pis Max 76.3858 +trainer/Log Pis Min -9.79102 +trainer/policy/mean Mean -0.031663 +trainer/policy/mean Std 0.906477 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86945 +trainer/policy/normal/std Std 0.702542 +trainer/policy/normal/std Max 6.89767 +trainer/policy/normal/std Min 0.254473 +trainer/policy/normal/log_std Mean 1.01028 +trainer/policy/normal/log_std Std 0.336341 +trainer/policy/normal/log_std Max 1.93118 +trainer/policy/normal/log_std Min -1.36856 +eval/num steps total 687715 +eval/num paths total 690 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0401375 +eval/Actions Std 0.901951 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61696 +time/logging (s) 0.00392174 +time/sampling batch (s) 0.529533 +time/saving (s) 0.00361405 +time/training (s) 7.40445 +time/epoch (s) 10.5585 +time/total (s) 7158.1 +Epoch -311 +---------------------------------- --------------- +2022-05-10 15:10:14.379612 PDT | [1] Epoch -310 finished +---------------------------------- --------------- +epoch -310 +replay_buffer/size 999033 +trainer/num train calls 691000 +trainer/Policy Loss -20.6136 +trainer/Log Pis Mean 24.7259 +trainer/Log Pis Std 14.0498 +trainer/Log Pis Max 73.1643 +trainer/Log Pis Min -8.92078 +trainer/policy/mean Mean -0.0255835 +trainer/policy/mean Std 0.908277 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82626 +trainer/policy/normal/std Std 0.700916 +trainer/policy/normal/std Max 5.48447 +trainer/policy/normal/std Min 0.229297 +trainer/policy/normal/log_std Mean 0.992566 +trainer/policy/normal/log_std Std 0.348674 +trainer/policy/normal/log_std Max 1.70192 +trainer/policy/normal/log_std Min -1.47274 +eval/num steps total 688715 +eval/num paths total 691 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0407142 +eval/Actions Std 0.918883 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64422 +time/logging (s) 0.00378615 +time/sampling batch (s) 0.279754 +time/saving (s) 0.00340984 +time/training (s) 7.00256 +time/epoch (s) 9.93373 +time/total (s) 7168.04 +Epoch -310 +---------------------------------- --------------- +2022-05-10 15:10:24.413822 PDT | [1] Epoch -309 finished +---------------------------------- --------------- +epoch -309 +replay_buffer/size 999033 +trainer/num train calls 692000 +trainer/Policy Loss -20.0226 +trainer/Log Pis Mean 24.4189 +trainer/Log Pis Std 13.3152 +trainer/Log Pis Max 78.854 +trainer/Log Pis Min -10.3105 +trainer/policy/mean Mean -0.0448271 +trainer/policy/mean Std 0.907261 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79963 +trainer/policy/normal/std Std 0.648033 +trainer/policy/normal/std Max 5.59025 +trainer/policy/normal/std Min 0.301285 +trainer/policy/normal/log_std Mean 0.991799 +trainer/policy/normal/log_std Std 0.307074 +trainer/policy/normal/log_std Max 1.72102 +trainer/policy/normal/log_std Min -1.1997 +eval/num steps total 689715 +eval/num paths total 692 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.111071 +eval/Actions Std 0.883054 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62713 +time/logging (s) 0.00376767 +time/sampling batch (s) 0.531134 +time/saving (s) 0.00344994 +time/training (s) 6.84466 +time/epoch (s) 10.0101 +time/total (s) 7178.05 +Epoch -309 +---------------------------------- --------------- +2022-05-10 15:10:33.974744 PDT | [1] Epoch -308 finished +---------------------------------- --------------- +epoch -308 +replay_buffer/size 999033 +trainer/num train calls 693000 +trainer/Policy Loss -20.4551 +trainer/Log Pis Mean 24.5998 +trainer/Log Pis Std 13.4271 +trainer/Log Pis Max 66.2863 +trainer/Log Pis Min -11.9377 +trainer/policy/mean Mean -0.0344359 +trainer/policy/mean Std 0.914234 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82973 +trainer/policy/normal/std Std 0.666091 +trainer/policy/normal/std Max 5.4663 +trainer/policy/normal/std Min 0.210934 +trainer/policy/normal/log_std Mean 0.999033 +trainer/policy/normal/log_std Std 0.326927 +trainer/policy/normal/log_std Max 1.6986 +trainer/policy/normal/log_std Min -1.55621 +eval/num steps total 690715 +eval/num paths total 693 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0948133 +eval/Actions Std 0.911085 +eval/Actions Max 0.99998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39124 +time/logging (s) 0.0049115 +time/sampling batch (s) 0.277537 +time/saving (s) 0.00346029 +time/training (s) 6.86124 +time/epoch (s) 9.53838 +time/total (s) 7187.59 +Epoch -308 +---------------------------------- --------------- +2022-05-10 15:10:44.028715 PDT | [1] Epoch -307 finished +---------------------------------- --------------- +epoch -307 +replay_buffer/size 999033 +trainer/num train calls 694000 +trainer/Policy Loss -18.9027 +trainer/Log Pis Mean 24.0951 +trainer/Log Pis Std 13.7075 +trainer/Log Pis Max 73.7143 +trainer/Log Pis Min -9.23828 +trainer/policy/mean Mean -0.0497647 +trainer/policy/mean Std 0.902499 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83793 +trainer/policy/normal/std Std 0.676765 +trainer/policy/normal/std Max 5.93525 +trainer/policy/normal/std Min 0.283422 +trainer/policy/normal/log_std Mean 1.00336 +trainer/policy/normal/log_std Std 0.316151 +trainer/policy/normal/log_std Max 1.78091 +trainer/policy/normal/log_std Min -1.26082 +eval/num steps total 691715 +eval/num paths total 694 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.418942 +eval/Actions Std 0.718155 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67608 +time/logging (s) 0.00424361 +time/sampling batch (s) 0.278148 +time/saving (s) 0.00384482 +time/training (s) 7.06735 +time/epoch (s) 10.0297 +time/total (s) 7197.62 +Epoch -307 +---------------------------------- --------------- +2022-05-10 15:10:54.529146 PDT | [1] Epoch -306 finished +---------------------------------- --------------- +epoch -306 +replay_buffer/size 999033 +trainer/num train calls 695000 +trainer/Policy Loss -19.3977 +trainer/Log Pis Mean 24.9803 +trainer/Log Pis Std 13.31 +trainer/Log Pis Max 65.2383 +trainer/Log Pis Min -9.71389 +trainer/policy/mean Mean -0.0385909 +trainer/policy/mean Std 0.906945 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80797 +trainer/policy/normal/std Std 0.686683 +trainer/policy/normal/std Max 6.49079 +trainer/policy/normal/std Min 0.242158 +trainer/policy/normal/log_std Mean 0.988609 +trainer/policy/normal/log_std Std 0.337352 +trainer/policy/normal/log_std Max 1.87038 +trainer/policy/normal/log_std Min -1.41816 +eval/num steps total 692715 +eval/num paths total 695 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0728655 +eval/Actions Std 0.912791 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51881 +time/logging (s) 0.00386104 +time/sampling batch (s) 0.776858 +time/saving (s) 0.00378736 +time/training (s) 7.17263 +time/epoch (s) 10.4759 +time/total (s) 7208.1 +Epoch -306 +---------------------------------- --------------- +2022-05-10 15:11:04.917028 PDT | [1] Epoch -305 finished +---------------------------------- --------------- +epoch -305 +replay_buffer/size 999033 +trainer/num train calls 696000 +trainer/Policy Loss -19.1892 +trainer/Log Pis Mean 24.9105 +trainer/Log Pis Std 13.3651 +trainer/Log Pis Max 81.5722 +trainer/Log Pis Min -7.65752 +trainer/policy/mean Mean -0.0465634 +trainer/policy/mean Std 0.90843 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82983 +trainer/policy/normal/std Std 0.657744 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.28917 +trainer/policy/normal/log_std Mean 1.00244 +trainer/policy/normal/log_std Std 0.30839 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24074 +eval/num steps total 693715 +eval/num paths total 696 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0940189 +eval/Actions Std 0.909484 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54446 +time/logging (s) 0.00376576 +time/sampling batch (s) 0.527608 +time/saving (s) 0.00342774 +time/training (s) 7.28492 +time/epoch (s) 10.3642 +time/total (s) 7218.47 +Epoch -305 +---------------------------------- --------------- +2022-05-10 15:11:15.119913 PDT | [1] Epoch -304 finished +---------------------------------- --------------- +epoch -304 +replay_buffer/size 999033 +trainer/num train calls 697000 +trainer/Policy Loss -20.607 +trainer/Log Pis Mean 23.4077 +trainer/Log Pis Std 13.3022 +trainer/Log Pis Max 66.1674 +trainer/Log Pis Min -7.3246 +trainer/policy/mean Mean -0.0190993 +trainer/policy/mean Std 0.907556 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80996 +trainer/policy/normal/std Std 0.661145 +trainer/policy/normal/std Max 5.79505 +trainer/policy/normal/std Min 0.282784 +trainer/policy/normal/log_std Mean 0.994425 +trainer/policy/normal/log_std Std 0.311947 +trainer/policy/normal/log_std Max 1.757 +trainer/policy/normal/log_std Min -1.26307 +eval/num steps total 694715 +eval/num paths total 697 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0161686 +eval/Actions Std 0.905332 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.25203 +time/logging (s) 0.003787 +time/sampling batch (s) 0.526716 +time/saving (s) 0.00340115 +time/training (s) 7.39347 +time/epoch (s) 10.1794 +time/total (s) 7228.65 +Epoch -304 +---------------------------------- --------------- +2022-05-10 15:11:25.169846 PDT | [1] Epoch -303 finished +---------------------------------- --------------- +epoch -303 +replay_buffer/size 999033 +trainer/num train calls 698000 +trainer/Policy Loss -18.7939 +trainer/Log Pis Mean 24.0506 +trainer/Log Pis Std 12.5472 +trainer/Log Pis Max 63.1695 +trainer/Log Pis Min -8.04632 +trainer/policy/mean Mean -0.0261346 +trainer/policy/mean Std 0.903167 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.76889 +trainer/policy/normal/std Std 0.659853 +trainer/policy/normal/std Max 5.40325 +trainer/policy/normal/std Min 0.315559 +trainer/policy/normal/log_std Mean 0.97857 +trainer/policy/normal/log_std Std 0.316016 +trainer/policy/normal/log_std Max 1.687 +trainer/policy/normal/log_std Min -1.15341 +eval/num steps total 695715 +eval/num paths total 698 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0214053 +eval/Actions Std 0.908722 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65028 +time/logging (s) 0.00424616 +time/sampling batch (s) 0.528225 +time/saving (s) 0.00399734 +time/training (s) 6.84004 +time/epoch (s) 10.0268 +time/total (s) 7238.68 +Epoch -303 +---------------------------------- --------------- +2022-05-10 15:11:35.966874 PDT | [1] Epoch -302 finished +---------------------------------- --------------- +epoch -302 +replay_buffer/size 999033 +trainer/num train calls 699000 +trainer/Policy Loss -20.0091 +trainer/Log Pis Mean 23.8374 +trainer/Log Pis Std 12.2159 +trainer/Log Pis Max 61.8984 +trainer/Log Pis Min -6.41834 +trainer/policy/mean Mean -0.0330861 +trainer/policy/mean Std 0.902634 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.74761 +trainer/policy/normal/std Std 0.656085 +trainer/policy/normal/std Max 6.00752 +trainer/policy/normal/std Min 0.275304 +trainer/policy/normal/log_std Mean 0.969074 +trainer/policy/normal/log_std Std 0.327748 +trainer/policy/normal/log_std Max 1.79301 +trainer/policy/normal/log_std Min -1.28988 +eval/num steps total 696715 +eval/num paths total 699 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0223654 +eval/Actions Std 0.907661 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56106 +time/logging (s) 0.00376338 +time/sampling batch (s) 0.274655 +time/saving (s) 0.00356457 +time/training (s) 7.92968 +time/epoch (s) 10.7727 +time/total (s) 7249.46 +Epoch -302 +---------------------------------- --------------- +2022-05-10 15:11:45.807963 PDT | [1] Epoch -301 finished +---------------------------------- --------------- +epoch -301 +replay_buffer/size 999033 +trainer/num train calls 700000 +trainer/Policy Loss -20.8253 +trainer/Log Pis Mean 24.793 +trainer/Log Pis Std 13.4704 +trainer/Log Pis Max 72.3747 +trainer/Log Pis Min -5.41171 +trainer/policy/mean Mean -0.0240104 +trainer/policy/mean Std 0.907647 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82487 +trainer/policy/normal/std Std 0.649033 +trainer/policy/normal/std Max 5.62623 +trainer/policy/normal/std Min 0.280765 +trainer/policy/normal/log_std Mean 1.00072 +trainer/policy/normal/log_std Std 0.30935 +trainer/policy/normal/log_std Max 1.72744 +trainer/policy/normal/log_std Min -1.27024 +eval/num steps total 697715 +eval/num paths total 700 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.310046 +eval/Actions Std 0.8747 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.2682 +time/logging (s) 0.00419877 +time/sampling batch (s) 0.27747 +time/saving (s) 0.0068168 +time/training (s) 7.26126 +time/epoch (s) 9.81794 +time/total (s) 7259.28 +Epoch -301 +---------------------------------- --------------- +2022-05-10 15:11:56.844841 PDT | [1] Epoch -300 finished +---------------------------------- --------------- +epoch -300 +replay_buffer/size 999033 +trainer/num train calls 701000 +trainer/Policy Loss -21.0981 +trainer/Log Pis Mean 23.5472 +trainer/Log Pis Std 13.1757 +trainer/Log Pis Max 66.3896 +trainer/Log Pis Min -7.97616 +trainer/policy/mean Mean -0.0248436 +trainer/policy/mean Std 0.904952 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.76459 +trainer/policy/normal/std Std 0.653443 +trainer/policy/normal/std Max 5.92977 +trainer/policy/normal/std Min 0.266265 +trainer/policy/normal/log_std Mean 0.974236 +trainer/policy/normal/log_std Std 0.336075 +trainer/policy/normal/log_std Max 1.77999 +trainer/policy/normal/log_std Min -1.32326 +eval/num steps total 698715 +eval/num paths total 701 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121731 +eval/Actions Std 0.911815 +eval/Actions Max 0.999994 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55665 +time/logging (s) 0.0040377 +time/sampling batch (s) 0.276591 +time/saving (s) 0.00397157 +time/training (s) 8.17162 +time/epoch (s) 11.0129 +time/total (s) 7270.3 +Epoch -300 +---------------------------------- --------------- +2022-05-10 15:12:06.470513 PDT | [1] Epoch -299 finished +---------------------------------- --------------- +epoch -299 +replay_buffer/size 999033 +trainer/num train calls 702000 +trainer/Policy Loss -20.5356 +trainer/Log Pis Mean 24.536 +trainer/Log Pis Std 13.6065 +trainer/Log Pis Max 71.5577 +trainer/Log Pis Min -10.5704 +trainer/policy/mean Mean -0.0208888 +trainer/policy/mean Std 0.909514 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85146 +trainer/policy/normal/std Std 0.697989 +trainer/policy/normal/std Max 6.55098 +trainer/policy/normal/std Min 0.244746 +trainer/policy/normal/log_std Mean 1.00424 +trainer/policy/normal/log_std Std 0.335531 +trainer/policy/normal/log_std Max 1.87961 +trainer/policy/normal/log_std Min -1.40753 +eval/num steps total 699715 +eval/num paths total 702 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0165755 +eval/Actions Std 0.90734 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83454 +time/logging (s) 0.00406931 +time/sampling batch (s) 0.526766 +time/saving (s) 0.00394226 +time/training (s) 6.2323 +time/epoch (s) 9.60161 +time/total (s) 7279.9 +Epoch -299 +---------------------------------- --------------- +2022-05-10 15:12:17.160054 PDT | [1] Epoch -298 finished +---------------------------------- --------------- +epoch -298 +replay_buffer/size 999033 +trainer/num train calls 703000 +trainer/Policy Loss -19.3371 +trainer/Log Pis Mean 23.5225 +trainer/Log Pis Std 12.9954 +trainer/Log Pis Max 67.3621 +trainer/Log Pis Min -7.2467 +trainer/policy/mean Mean -0.0506939 +trainer/policy/mean Std 0.905298 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77742 +trainer/policy/normal/std Std 0.658562 +trainer/policy/normal/std Max 5.91707 +trainer/policy/normal/std Min 0.280866 +trainer/policy/normal/log_std Mean 0.979714 +trainer/policy/normal/log_std Std 0.331055 +trainer/policy/normal/log_std Max 1.77784 +trainer/policy/normal/log_std Min -1.26988 +eval/num steps total 700715 +eval/num paths total 703 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.195808 +eval/Actions Std 0.954028 +eval/Actions Max 0.999984 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56447 +time/logging (s) 0.00412355 +time/sampling batch (s) 0.530396 +time/saving (s) 0.00395081 +time/training (s) 7.56277 +time/epoch (s) 10.6657 +time/total (s) 7290.57 +Epoch -298 +---------------------------------- --------------- +2022-05-10 15:12:27.146271 PDT | [1] Epoch -297 finished +---------------------------------- --------------- +epoch -297 +replay_buffer/size 999033 +trainer/num train calls 704000 +trainer/Policy Loss -20.2269 +trainer/Log Pis Mean 23.5295 +trainer/Log Pis Std 13.0878 +trainer/Log Pis Max 67.1566 +trainer/Log Pis Min -10.2612 +trainer/policy/mean Mean -0.0331573 +trainer/policy/mean Std 0.907892 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76929 +trainer/policy/normal/std Std 0.656478 +trainer/policy/normal/std Max 5.41827 +trainer/policy/normal/std Min 0.262449 +trainer/policy/normal/log_std Mean 0.977563 +trainer/policy/normal/log_std Std 0.324753 +trainer/policy/normal/log_std Max 1.68978 +trainer/policy/normal/log_std Min -1.3377 +eval/num steps total 701715 +eval/num paths total 704 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0171898 +eval/Actions Std 0.907274 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50702 +time/logging (s) 0.00410598 +time/sampling batch (s) 0.528666 +time/saving (s) 0.00386669 +time/training (s) 6.91823 +time/epoch (s) 9.96188 +time/total (s) 7300.53 +Epoch -297 +---------------------------------- --------------- +2022-05-10 15:12:37.676886 PDT | [1] Epoch -296 finished +---------------------------------- --------------- +epoch -296 +replay_buffer/size 999033 +trainer/num train calls 705000 +trainer/Policy Loss -20.4101 +trainer/Log Pis Mean 25.3307 +trainer/Log Pis Std 12.8743 +trainer/Log Pis Max 61.9895 +trainer/Log Pis Min -5.26836 +trainer/policy/mean Mean -0.0426312 +trainer/policy/mean Std 0.906949 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8682 +trainer/policy/normal/std Std 0.66986 +trainer/policy/normal/std Max 6.1357 +trainer/policy/normal/std Min 0.230981 +trainer/policy/normal/log_std Mean 1.01497 +trainer/policy/normal/log_std Std 0.313818 +trainer/policy/normal/log_std Max 1.81413 +trainer/policy/normal/log_std Min -1.46542 +eval/num steps total 702715 +eval/num paths total 705 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.453269 +eval/Actions Std 0.810861 +eval/Actions Max 0.999985 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64612 +time/logging (s) 0.00382092 +time/sampling batch (s) 0.529324 +time/saving (s) 0.00358268 +time/training (s) 7.32334 +time/epoch (s) 10.5062 +time/total (s) 7311.04 +Epoch -296 +---------------------------------- --------------- +2022-05-10 15:12:47.528573 PDT | [1] Epoch -295 finished +---------------------------------- --------------- +epoch -295 +replay_buffer/size 999033 +trainer/num train calls 706000 +trainer/Policy Loss -18.6281 +trainer/Log Pis Mean 24.3757 +trainer/Log Pis Std 13.6249 +trainer/Log Pis Max 75.5101 +trainer/Log Pis Min -7.25561 +trainer/policy/mean Mean -0.0442235 +trainer/policy/mean Std 0.903988 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80155 +trainer/policy/normal/std Std 0.634486 +trainer/policy/normal/std Max 5.55698 +trainer/policy/normal/std Min 0.314001 +trainer/policy/normal/log_std Mean 0.995439 +trainer/policy/normal/log_std Std 0.292572 +trainer/policy/normal/log_std Max 1.71505 +trainer/policy/normal/log_std Min -1.15836 +eval/num steps total 703715 +eval/num paths total 706 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.212832 +eval/Actions Std 0.879557 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51777 +time/logging (s) 0.00411565 +time/sampling batch (s) 0.278709 +time/saving (s) 0.00394926 +time/training (s) 7.02347 +time/epoch (s) 9.828 +time/total (s) 7320.87 +Epoch -295 +---------------------------------- --------------- +2022-05-10 15:12:58.129738 PDT | [1] Epoch -294 finished +---------------------------------- --------------- +epoch -294 +replay_buffer/size 999033 +trainer/num train calls 707000 +trainer/Policy Loss -20.3617 +trainer/Log Pis Mean 24.5137 +trainer/Log Pis Std 13.0901 +trainer/Log Pis Max 71.9229 +trainer/Log Pis Min -9.10744 +trainer/policy/mean Mean -0.0562971 +trainer/policy/mean Std 0.910881 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75346 +trainer/policy/normal/std Std 0.64121 +trainer/policy/normal/std Max 5.04807 +trainer/policy/normal/std Min 0.280903 +trainer/policy/normal/log_std Mean 0.973287 +trainer/policy/normal/log_std Std 0.318339 +trainer/policy/normal/log_std Max 1.61901 +trainer/policy/normal/log_std Min -1.26974 +eval/num steps total 704715 +eval/num paths total 707 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.156066 +eval/Actions Std 0.941864 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89338 +time/logging (s) 0.00376737 +time/sampling batch (s) 0.278409 +time/saving (s) 0.00361231 +time/training (s) 7.3974 +time/epoch (s) 10.5766 +time/total (s) 7331.45 +Epoch -294 +---------------------------------- --------------- +2022-05-10 15:13:08.424177 PDT | [1] Epoch -293 finished +---------------------------------- --------------- +epoch -293 +replay_buffer/size 999033 +trainer/num train calls 708000 +trainer/Policy Loss -20.2481 +trainer/Log Pis Mean 24.4141 +trainer/Log Pis Std 13.6837 +trainer/Log Pis Max 67.8416 +trainer/Log Pis Min -8.48113 +trainer/policy/mean Mean -0.0415579 +trainer/policy/mean Std 0.910018 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82473 +trainer/policy/normal/std Std 0.653058 +trainer/policy/normal/std Max 5.9382 +trainer/policy/normal/std Min 0.28632 +trainer/policy/normal/log_std Mean 1.00022 +trainer/policy/normal/log_std Std 0.310918 +trainer/policy/normal/log_std Max 1.78141 +trainer/policy/normal/log_std Min -1.25064 +eval/num steps total 705715 +eval/num paths total 708 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.2548 +eval/Actions Std 0.840934 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57876 +time/logging (s) 0.00398587 +time/sampling batch (s) 0.526823 +time/saving (s) 0.00353535 +time/training (s) 7.15787 +time/epoch (s) 10.271 +time/total (s) 7341.73 +Epoch -293 +---------------------------------- --------------- +2022-05-10 15:13:18.604947 PDT | [1] Epoch -292 finished +---------------------------------- -------------- +epoch -292 +replay_buffer/size 999033 +trainer/num train calls 709000 +trainer/Policy Loss -19.3083 +trainer/Log Pis Mean 24.4614 +trainer/Log Pis Std 13.6047 +trainer/Log Pis Max 72.1484 +trainer/Log Pis Min -8.09617 +trainer/policy/mean Mean -0.0272414 +trainer/policy/mean Std 0.906968 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81294 +trainer/policy/normal/std Std 0.650968 +trainer/policy/normal/std Max 6.29657 +trainer/policy/normal/std Min 0.347439 +trainer/policy/normal/log_std Mean 0.996832 +trainer/policy/normal/log_std Std 0.306332 +trainer/policy/normal/log_std Max 1.84001 +trainer/policy/normal/log_std Min -1.05717 +eval/num steps total 706715 +eval/num paths total 709 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135557 +eval/Actions Std 0.882432 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.24277 +time/logging (s) 0.0063598 +time/sampling batch (s) 0.292689 +time/saving (s) 0.0053196 +time/training (s) 7.61151 +time/epoch (s) 10.1586 +time/total (s) 7351.89 +Epoch -292 +---------------------------------- -------------- +2022-05-10 15:13:28.656699 PDT | [1] Epoch -291 finished +---------------------------------- --------------- +epoch -291 +replay_buffer/size 999033 +trainer/num train calls 710000 +trainer/Policy Loss -21.0627 +trainer/Log Pis Mean 24.7575 +trainer/Log Pis Std 12.7107 +trainer/Log Pis Max 64.9675 +trainer/Log Pis Min -8.17053 +trainer/policy/mean Mean -0.0441866 +trainer/policy/mean Std 0.91062 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77662 +trainer/policy/normal/std Std 0.645918 +trainer/policy/normal/std Max 5.21429 +trainer/policy/normal/std Min 0.300845 +trainer/policy/normal/log_std Mean 0.981899 +trainer/policy/normal/log_std Std 0.317386 +trainer/policy/normal/log_std Max 1.6514 +trainer/policy/normal/log_std Min -1.20116 +eval/num steps total 707715 +eval/num paths total 710 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0289622 +eval/Actions Std 0.908269 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89475 +time/logging (s) 0.00385773 +time/sampling batch (s) 0.288235 +time/saving (s) 0.00357698 +time/training (s) 6.83097 +time/epoch (s) 10.0214 +time/total (s) 7361.92 +Epoch -291 +---------------------------------- --------------- +2022-05-10 15:13:39.315196 PDT | [1] Epoch -290 finished +---------------------------------- --------------- +epoch -290 +replay_buffer/size 999033 +trainer/num train calls 711000 +trainer/Policy Loss -20.3589 +trainer/Log Pis Mean 24.9651 +trainer/Log Pis Std 13.3955 +trainer/Log Pis Max 75.4032 +trainer/Log Pis Min -8.56755 +trainer/policy/mean Mean -0.0332078 +trainer/policy/mean Std 0.908058 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84318 +trainer/policy/normal/std Std 0.665256 +trainer/policy/normal/std Max 6.29277 +trainer/policy/normal/std Min 0.277176 +trainer/policy/normal/log_std Mean 1.00568 +trainer/policy/normal/log_std Std 0.31606 +trainer/policy/normal/log_std Max 1.8394 +trainer/policy/normal/log_std Min -1.2831 +eval/num steps total 708715 +eval/num paths total 711 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.435827 +eval/Actions Std 0.886145 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85543 +time/logging (s) 0.00415432 +time/sampling batch (s) 0.277089 +time/saving (s) 0.00348757 +time/training (s) 7.49497 +time/epoch (s) 10.6351 +time/total (s) 7372.55 +Epoch -290 +---------------------------------- --------------- +2022-05-10 15:13:50.886830 PDT | [1] Epoch -289 finished +---------------------------------- --------------- +epoch -289 +replay_buffer/size 999033 +trainer/num train calls 712000 +trainer/Policy Loss -18.9721 +trainer/Log Pis Mean 24.9527 +trainer/Log Pis Std 13.7094 +trainer/Log Pis Max 74.3316 +trainer/Log Pis Min -5.99543 +trainer/policy/mean Mean -0.0367218 +trainer/policy/mean Std 0.908129 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81595 +trainer/policy/normal/std Std 0.671883 +trainer/policy/normal/std Max 5.6002 +trainer/policy/normal/std Min 0.221905 +trainer/policy/normal/log_std Mean 0.993189 +trainer/policy/normal/log_std Std 0.331583 +trainer/policy/normal/log_std Max 1.7228 +trainer/policy/normal/log_std Min -1.5055 +eval/num steps total 709715 +eval/num paths total 712 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0746045 +eval/Actions Std 0.912978 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74448 +time/logging (s) 0.00422033 +time/sampling batch (s) 0.281814 +time/saving (s) 0.00410296 +time/training (s) 8.51112 +time/epoch (s) 11.5457 +time/total (s) 7384.1 +Epoch -289 +---------------------------------- --------------- +2022-05-10 15:14:01.720270 PDT | [1] Epoch -288 finished +---------------------------------- --------------- +epoch -288 +replay_buffer/size 999033 +trainer/num train calls 713000 +trainer/Policy Loss -20.1634 +trainer/Log Pis Mean 23.5799 +trainer/Log Pis Std 13.276 +trainer/Log Pis Max 61.941 +trainer/Log Pis Min -9.41387 +trainer/policy/mean Mean -0.0367205 +trainer/policy/mean Std 0.91004 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.73265 +trainer/policy/normal/std Std 0.642577 +trainer/policy/normal/std Max 5.47554 +trainer/policy/normal/std Min 0.240413 +trainer/policy/normal/log_std Mean 0.965201 +trainer/policy/normal/log_std Std 0.319806 +trainer/policy/normal/log_std Max 1.70029 +trainer/policy/normal/log_std Min -1.4254 +eval/num steps total 710715 +eval/num paths total 713 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00735436 +eval/Actions Std 0.905372 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48722 +time/logging (s) 0.00400645 +time/sampling batch (s) 0.530433 +time/saving (s) 0.00356437 +time/training (s) 7.78346 +time/epoch (s) 10.8087 +time/total (s) 7394.92 +Epoch -288 +---------------------------------- --------------- +2022-05-10 15:14:12.107060 PDT | [1] Epoch -287 finished +---------------------------------- --------------- +epoch -287 +replay_buffer/size 999033 +trainer/num train calls 714000 +trainer/Policy Loss -19.7188 +trainer/Log Pis Mean 24.6744 +trainer/Log Pis Std 13.6645 +trainer/Log Pis Max 65.0289 +trainer/Log Pis Min -4.57134 +trainer/policy/mean Mean -0.0277364 +trainer/policy/mean Std 0.904414 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81859 +trainer/policy/normal/std Std 0.671112 +trainer/policy/normal/std Max 6.63943 +trainer/policy/normal/std Min 0.257882 +trainer/policy/normal/log_std Mean 0.995354 +trainer/policy/normal/log_std Std 0.32326 +trainer/policy/normal/log_std Max 1.89303 +trainer/policy/normal/log_std Min -1.35525 +eval/num steps total 711715 +eval/num paths total 714 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.187325 +eval/Actions Std 0.899195 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55486 +time/logging (s) 0.00366842 +time/sampling batch (s) 0.280246 +time/saving (s) 0.00351826 +time/training (s) 7.51992 +time/epoch (s) 10.3622 +time/total (s) 7405.28 +Epoch -287 +---------------------------------- --------------- +2022-05-10 15:14:22.536789 PDT | [1] Epoch -286 finished +---------------------------------- --------------- +epoch -286 +replay_buffer/size 999033 +trainer/num train calls 715000 +trainer/Policy Loss -20.416 +trainer/Log Pis Mean 23.9757 +trainer/Log Pis Std 12.7763 +trainer/Log Pis Max 64.5861 +trainer/Log Pis Min -5.34466 +trainer/policy/mean Mean -0.0115217 +trainer/policy/mean Std 0.908084 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.74374 +trainer/policy/normal/std Std 0.646406 +trainer/policy/normal/std Max 6.41731 +trainer/policy/normal/std Min 0.268925 +trainer/policy/normal/log_std Mean 0.970218 +trainer/policy/normal/log_std Std 0.314639 +trainer/policy/normal/log_std Max 1.859 +trainer/policy/normal/log_std Min -1.31332 +eval/num steps total 712715 +eval/num paths total 715 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0969477 +eval/Actions Std 0.918334 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48049 +time/logging (s) 0.00383622 +time/sampling batch (s) 0.279802 +time/saving (s) 0.00345628 +time/training (s) 7.63839 +time/epoch (s) 10.406 +time/total (s) 7415.69 +Epoch -286 +---------------------------------- --------------- +2022-05-10 15:14:33.075696 PDT | [1] Epoch -285 finished +---------------------------------- --------------- +epoch -285 +replay_buffer/size 999033 +trainer/num train calls 716000 +trainer/Policy Loss -19.4933 +trainer/Log Pis Mean 23.614 +trainer/Log Pis Std 12.6513 +trainer/Log Pis Max 57.7083 +trainer/Log Pis Min -6.31852 +trainer/policy/mean Mean -0.0353685 +trainer/policy/mean Std 0.901561 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79356 +trainer/policy/normal/std Std 0.668582 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.252215 +trainer/policy/normal/log_std Mean 0.984672 +trainer/policy/normal/log_std Std 0.333957 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.37747 +eval/num steps total 713715 +eval/num paths total 716 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0617483 +eval/Actions Std 0.916449 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59361 +time/logging (s) 0.004255 +time/sampling batch (s) 0.281346 +time/saving (s) 0.00408493 +time/training (s) 7.63202 +time/epoch (s) 10.5153 +time/total (s) 7426.21 +Epoch -285 +---------------------------------- --------------- +2022-05-10 15:14:43.087789 PDT | [1] Epoch -284 finished +---------------------------------- --------------- +epoch -284 +replay_buffer/size 999033 +trainer/num train calls 717000 +trainer/Policy Loss -20.9961 +trainer/Log Pis Mean 24.3775 +trainer/Log Pis Std 12.79 +trainer/Log Pis Max 72.1948 +trainer/Log Pis Min -8.24807 +trainer/policy/mean Mean -0.0381141 +trainer/policy/mean Std 0.912284 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76488 +trainer/policy/normal/std Std 0.6642 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.256931 +trainer/policy/normal/log_std Mean 0.975001 +trainer/policy/normal/log_std Std 0.328196 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.35895 +eval/num steps total 714715 +eval/num paths total 717 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0130315 +eval/Actions Std 0.90338 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49699 +time/logging (s) 0.00376862 +time/sampling batch (s) 0.28135 +time/saving (s) 0.00356495 +time/training (s) 7.20144 +time/epoch (s) 9.98711 +time/total (s) 7436.2 +Epoch -284 +---------------------------------- --------------- +2022-05-10 15:14:52.410975 PDT | [1] Epoch -283 finished +---------------------------------- --------------- +epoch -283 +replay_buffer/size 999033 +trainer/num train calls 718000 +trainer/Policy Loss -19.0394 +trainer/Log Pis Mean 24.6657 +trainer/Log Pis Std 14.0718 +trainer/Log Pis Max 69.9666 +trainer/Log Pis Min -8.68151 +trainer/policy/mean Mean -0.0372249 +trainer/policy/mean Std 0.905497 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83955 +trainer/policy/normal/std Std 0.675832 +trainer/policy/normal/std Max 6.32076 +trainer/policy/normal/std Min 0.289467 +trainer/policy/normal/log_std Mean 1.0021 +trainer/policy/normal/log_std Std 0.327844 +trainer/policy/normal/log_std Max 1.84384 +trainer/policy/normal/log_std Min -1.23971 +eval/num steps total 715715 +eval/num paths total 718 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0304978 +eval/Actions Std 0.90524 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38563 +time/logging (s) 0.00378408 +time/sampling batch (s) 0.303334 +time/saving (s) 0.0034704 +time/training (s) 6.60286 +time/epoch (s) 9.29908 +time/total (s) 7445.5 +Epoch -283 +---------------------------------- --------------- +2022-05-10 15:15:02.424580 PDT | [1] Epoch -282 finished +---------------------------------- --------------- +epoch -282 +replay_buffer/size 999033 +trainer/num train calls 719000 +trainer/Policy Loss -19.1649 +trainer/Log Pis Mean 24.8493 +trainer/Log Pis Std 13.5035 +trainer/Log Pis Max 72.7795 +trainer/Log Pis Min -7.90781 +trainer/policy/mean Mean -0.04064 +trainer/policy/mean Std 0.906788 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81438 +trainer/policy/normal/std Std 0.685389 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.254979 +trainer/policy/normal/log_std Mean 0.990573 +trainer/policy/normal/log_std Std 0.339064 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.36657 +eval/num steps total 716715 +eval/num paths total 719 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0192598 +eval/Actions Std 0.905205 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.90062 +time/logging (s) 0.00367198 +time/sampling batch (s) 0.280386 +time/saving (s) 0.00343499 +time/training (s) 6.80113 +time/epoch (s) 9.98925 +time/total (s) 7455.49 +Epoch -282 +---------------------------------- --------------- +2022-05-10 15:15:11.548104 PDT | [1] Epoch -281 finished +---------------------------------- --------------- +epoch -281 +replay_buffer/size 999033 +trainer/num train calls 720000 +trainer/Policy Loss -19.5635 +trainer/Log Pis Mean 24.1804 +trainer/Log Pis Std 12.9954 +trainer/Log Pis Max 67.3301 +trainer/Log Pis Min -7.20403 +trainer/policy/mean Mean -0.0276192 +trainer/policy/mean Std 0.906366 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82225 +trainer/policy/normal/std Std 0.656348 +trainer/policy/normal/std Max 6.64675 +trainer/policy/normal/std Min 0.316452 +trainer/policy/normal/log_std Mean 0.998685 +trainer/policy/normal/log_std Std 0.315221 +trainer/policy/normal/log_std Max 1.89413 +trainer/policy/normal/log_std Min -1.15058 +eval/num steps total 717715 +eval/num paths total 720 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0136275 +eval/Actions Std 0.912915 +eval/Actions Max 0.999993 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36731 +time/logging (s) 0.00420607 +time/sampling batch (s) 0.278696 +time/saving (s) 0.00381575 +time/training (s) 6.44615 +time/epoch (s) 9.10018 +time/total (s) 7464.6 +Epoch -281 +---------------------------------- --------------- +2022-05-10 15:15:21.198738 PDT | [1] Epoch -280 finished +---------------------------------- --------------- +epoch -280 +replay_buffer/size 999033 +trainer/num train calls 721000 +trainer/Policy Loss -20.0204 +trainer/Log Pis Mean 23.8863 +trainer/Log Pis Std 13.4052 +trainer/Log Pis Max 67.6302 +trainer/Log Pis Min -9.62104 +trainer/policy/mean Mean -0.0165947 +trainer/policy/mean Std 0.906236 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76761 +trainer/policy/normal/std Std 0.671498 +trainer/policy/normal/std Max 5.66812 +trainer/policy/normal/std Min 0.293862 +trainer/policy/normal/log_std Mean 0.974304 +trainer/policy/normal/log_std Std 0.336074 +trainer/policy/normal/log_std Max 1.73486 +trainer/policy/normal/log_std Min -1.22465 +eval/num steps total 718715 +eval/num paths total 721 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.090273 +eval/Actions Std 0.875238 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53782 +time/logging (s) 0.00387436 +time/sampling batch (s) 0.278673 +time/saving (s) 0.00374943 +time/training (s) 6.80183 +time/epoch (s) 9.62595 +time/total (s) 7474.23 +Epoch -280 +---------------------------------- --------------- +2022-05-10 15:15:32.738908 PDT | [1] Epoch -279 finished +---------------------------------- --------------- +epoch -279 +replay_buffer/size 999033 +trainer/num train calls 722000 +trainer/Policy Loss -19.286 +trainer/Log Pis Mean 24.235 +trainer/Log Pis Std 12.9228 +trainer/Log Pis Max 65.6805 +trainer/Log Pis Min -6.72909 +trainer/policy/mean Mean -0.0311959 +trainer/policy/mean Std 0.906405 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8007 +trainer/policy/normal/std Std 0.686094 +trainer/policy/normal/std Max 6.8978 +trainer/policy/normal/std Min 0.301029 +trainer/policy/normal/log_std Mean 0.985584 +trainer/policy/normal/log_std Std 0.338653 +trainer/policy/normal/log_std Max 1.9312 +trainer/policy/normal/log_std Min -1.20055 +eval/num steps total 719715 +eval/num paths total 722 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125432 +eval/Actions Std 0.90951 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68731 +time/logging (s) 0.00382972 +time/sampling batch (s) 1.02983 +time/saving (s) 0.00342618 +time/training (s) 7.79166 +time/epoch (s) 11.516 +time/total (s) 7485.74 +Epoch -279 +---------------------------------- --------------- +2022-05-10 15:15:42.670444 PDT | [1] Epoch -278 finished +---------------------------------- --------------- +epoch -278 +replay_buffer/size 999033 +trainer/num train calls 723000 +trainer/Policy Loss -20.5323 +trainer/Log Pis Mean 25.2547 +trainer/Log Pis Std 14.102 +trainer/Log Pis Max 86.2404 +trainer/Log Pis Min -5.03908 +trainer/policy/mean Mean -0.0289652 +trainer/policy/mean Std 0.909206 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.80008 +trainer/policy/normal/std Std 0.673 +trainer/policy/normal/std Max 6.40617 +trainer/policy/normal/std Min 0.2742 +trainer/policy/normal/log_std Mean 0.988172 +trainer/policy/normal/log_std Std 0.325302 +trainer/policy/normal/log_std Max 1.85726 +trainer/policy/normal/log_std Min -1.2939 +eval/num steps total 720715 +eval/num paths total 723 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00150707 +eval/Actions Std 0.923408 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33663 +time/logging (s) 0.00389847 +time/sampling batch (s) 0.527885 +time/saving (s) 0.00345167 +time/training (s) 7.03586 +time/epoch (s) 9.90772 +time/total (s) 7495.66 +Epoch -278 +---------------------------------- --------------- +2022-05-10 15:15:52.398086 PDT | [1] Epoch -277 finished +---------------------------------- --------------- +epoch -277 +replay_buffer/size 999033 +trainer/num train calls 724000 +trainer/Policy Loss -19.2516 +trainer/Log Pis Mean 24.19 +trainer/Log Pis Std 13.1618 +trainer/Log Pis Max 61.8788 +trainer/Log Pis Min -8.31451 +trainer/policy/mean Mean -0.0260948 +trainer/policy/mean Std 0.905565 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8099 +trainer/policy/normal/std Std 0.679482 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.320134 +trainer/policy/normal/log_std Mean 0.992843 +trainer/policy/normal/log_std Std 0.316596 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.13902 +eval/num steps total 721715 +eval/num paths total 724 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.210887 +eval/Actions Std 0.955618 +eval/Actions Max 0.999974 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.626 +time/logging (s) 0.00374457 +time/sampling batch (s) 0.279032 +time/saving (s) 0.00340608 +time/training (s) 6.79141 +time/epoch (s) 9.70359 +time/total (s) 7505.36 +Epoch -277 +---------------------------------- --------------- +2022-05-10 15:16:02.659164 PDT | [1] Epoch -276 finished +---------------------------------- --------------- +epoch -276 +replay_buffer/size 999033 +trainer/num train calls 725000 +trainer/Policy Loss -20.2807 +trainer/Log Pis Mean 25.3405 +trainer/Log Pis Std 13.6612 +trainer/Log Pis Max 69.8415 +trainer/Log Pis Min -3.56059 +trainer/policy/mean Mean -0.0276048 +trainer/policy/mean Std 0.908934 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8309 +trainer/policy/normal/std Std 0.663704 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.266606 +trainer/policy/normal/log_std Mean 1.00038 +trainer/policy/normal/log_std Std 0.32291 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.32198 +eval/num steps total 722715 +eval/num paths total 725 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0839213 +eval/Actions Std 0.890266 +eval/Actions Max 0.999996 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73435 +time/logging (s) 0.00431139 +time/sampling batch (s) 0.781224 +time/saving (s) 0.00413696 +time/training (s) 6.71356 +time/epoch (s) 10.2376 +time/total (s) 7515.6 +Epoch -276 +---------------------------------- --------------- +2022-05-10 15:16:13.426638 PDT | [1] Epoch -275 finished +---------------------------------- --------------- +epoch -275 +replay_buffer/size 999033 +trainer/num train calls 726000 +trainer/Policy Loss -19.9158 +trainer/Log Pis Mean 24.9771 +trainer/Log Pis Std 13.3804 +trainer/Log Pis Max 66.1735 +trainer/Log Pis Min -4.23717 +trainer/policy/mean Mean -0.0391509 +trainer/policy/mean Std 0.91195 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.85198 +trainer/policy/normal/std Std 0.64718 +trainer/policy/normal/std Max 6.30224 +trainer/policy/normal/std Min 0.275171 +trainer/policy/normal/log_std Mean 1.01338 +trainer/policy/normal/log_std Std 0.292192 +trainer/policy/normal/log_std Max 1.84091 +trainer/policy/normal/log_std Min -1.29036 +eval/num steps total 723715 +eval/num paths total 726 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.29147 +eval/Actions Std 0.843911 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60295 +time/logging (s) 0.00379843 +time/sampling batch (s) 0.280939 +time/saving (s) 0.00342807 +time/training (s) 7.85123 +time/epoch (s) 10.7423 +time/total (s) 7526.35 +Epoch -275 +---------------------------------- --------------- +2022-05-10 15:16:23.855915 PDT | [1] Epoch -274 finished +---------------------------------- --------------- +epoch -274 +replay_buffer/size 999033 +trainer/num train calls 727000 +trainer/Policy Loss -20.3868 +trainer/Log Pis Mean 25.4409 +trainer/Log Pis Std 12.8084 +trainer/Log Pis Max 77.6367 +trainer/Log Pis Min -9.57599 +trainer/policy/mean Mean -0.00827992 +trainer/policy/mean Std 0.911765 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82913 +trainer/policy/normal/std Std 0.688322 +trainer/policy/normal/std Max 6.39218 +trainer/policy/normal/std Min 0.263038 +trainer/policy/normal/log_std Mean 0.997207 +trainer/policy/normal/log_std Std 0.331467 +trainer/policy/normal/log_std Max 1.85508 +trainer/policy/normal/log_std Min -1.33546 +eval/num steps total 724715 +eval/num paths total 727 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.147259 +eval/Actions Std 0.938473 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61039 +time/logging (s) 0.00375355 +time/sampling batch (s) 0.303918 +time/saving (s) 0.00345183 +time/training (s) 7.48338 +time/epoch (s) 10.4049 +time/total (s) 7536.76 +Epoch -274 +---------------------------------- --------------- +2022-05-10 15:16:33.594308 PDT | [1] Epoch -273 finished +---------------------------------- --------------- +epoch -273 +replay_buffer/size 999033 +trainer/num train calls 728000 +trainer/Policy Loss -20.5949 +trainer/Log Pis Mean 25.2335 +trainer/Log Pis Std 13.2749 +trainer/Log Pis Max 68.0412 +trainer/Log Pis Min -7.48454 +trainer/policy/mean Mean -0.0433144 +trainer/policy/mean Std 0.90561 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82645 +trainer/policy/normal/std Std 0.668529 +trainer/policy/normal/std Max 6.86107 +trainer/policy/normal/std Min 0.329874 +trainer/policy/normal/log_std Mean 0.998172 +trainer/policy/normal/log_std Std 0.324371 +trainer/policy/normal/log_std Max 1.92586 +trainer/policy/normal/log_std Min -1.10904 +eval/num steps total 725715 +eval/num paths total 728 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.150926 +eval/Actions Std 0.877277 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72394 +time/logging (s) 0.0036915 +time/sampling batch (s) 0.282091 +time/saving (s) 0.00345227 +time/training (s) 6.70096 +time/epoch (s) 9.71413 +time/total (s) 7546.47 +Epoch -273 +---------------------------------- --------------- +2022-05-10 15:16:44.455522 PDT | [1] Epoch -272 finished +---------------------------------- --------------- +epoch -272 +replay_buffer/size 999033 +trainer/num train calls 729000 +trainer/Policy Loss -19.8091 +trainer/Log Pis Mean 24.4221 +trainer/Log Pis Std 13.3208 +trainer/Log Pis Max 63.3524 +trainer/Log Pis Min -8.88598 +trainer/policy/mean Mean -0.0477475 +trainer/policy/mean Std 0.9132 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81032 +trainer/policy/normal/std Std 0.632025 +trainer/policy/normal/std Max 6.73048 +trainer/policy/normal/std Min 0.319988 +trainer/policy/normal/log_std Mean 0.99796 +trainer/policy/normal/log_std Std 0.297895 +trainer/policy/normal/log_std Max 1.90665 +trainer/policy/normal/log_std Min -1.13947 +eval/num steps total 726715 +eval/num paths total 729 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00642185 +eval/Actions Std 0.909915 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5256 +time/logging (s) 0.00366472 +time/sampling batch (s) 0.532085 +time/saving (s) 0.00343071 +time/training (s) 7.77219 +time/epoch (s) 10.837 +time/total (s) 7557.31 +Epoch -272 +---------------------------------- --------------- +2022-05-10 15:16:55.027422 PDT | [1] Epoch -271 finished +---------------------------------- --------------- +epoch -271 +replay_buffer/size 999033 +trainer/num train calls 730000 +trainer/Policy Loss -19.3695 +trainer/Log Pis Mean 24.3307 +trainer/Log Pis Std 13.6581 +trainer/Log Pis Max 70.5474 +trainer/Log Pis Min -12.9721 +trainer/policy/mean Mean -0.0415784 +trainer/policy/mean Std 0.904239 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80963 +trainer/policy/normal/std Std 0.688366 +trainer/policy/normal/std Max 6.03525 +trainer/policy/normal/std Min 0.295647 +trainer/policy/normal/log_std Mean 0.989952 +trainer/policy/normal/log_std Std 0.331184 +trainer/policy/normal/log_std Max 1.79762 +trainer/policy/normal/log_std Min -1.21859 +eval/num steps total 727715 +eval/num paths total 730 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.31343 +eval/Actions Std 0.861603 +eval/Actions Max 0.999979 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52333 +time/logging (s) 0.00395373 +time/sampling batch (s) 0.280366 +time/saving (s) 0.00400204 +time/training (s) 7.73636 +time/epoch (s) 10.548 +time/total (s) 7567.86 +Epoch -271 +---------------------------------- --------------- +2022-05-10 15:17:04.676482 PDT | [1] Epoch -270 finished +---------------------------------- --------------- +epoch -270 +replay_buffer/size 999033 +trainer/num train calls 731000 +trainer/Policy Loss -20.0592 +trainer/Log Pis Mean 24.1221 +trainer/Log Pis Std 13.1576 +trainer/Log Pis Max 63.4799 +trainer/Log Pis Min -11.2721 +trainer/policy/mean Mean -0.0385568 +trainer/policy/mean Std 0.907924 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80067 +trainer/policy/normal/std Std 0.667706 +trainer/policy/normal/std Max 6.0004 +trainer/policy/normal/std Min 0.254422 +trainer/policy/normal/log_std Mean 0.987606 +trainer/policy/normal/log_std Std 0.331807 +trainer/policy/normal/log_std Max 1.79183 +trainer/policy/normal/log_std Min -1.36876 +eval/num steps total 728715 +eval/num paths total 731 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.091557 +eval/Actions Std 0.918427 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5351 +time/logging (s) 0.0037275 +time/sampling batch (s) 0.281299 +time/saving (s) 0.00342646 +time/training (s) 6.80076 +time/epoch (s) 9.62431 +time/total (s) 7577.49 +Epoch -270 +---------------------------------- --------------- +2022-05-10 15:17:15.078821 PDT | [1] Epoch -269 finished +---------------------------------- --------------- +epoch -269 +replay_buffer/size 999033 +trainer/num train calls 732000 +trainer/Policy Loss -19.3589 +trainer/Log Pis Mean 24.6507 +trainer/Log Pis Std 13.1865 +trainer/Log Pis Max 62.2859 +trainer/Log Pis Min -6.79952 +trainer/policy/mean Mean -0.052606 +trainer/policy/mean Std 0.904139 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.76328 +trainer/policy/normal/std Std 0.659102 +trainer/policy/normal/std Max 5.32363 +trainer/policy/normal/std Min 0.29876 +trainer/policy/normal/log_std Mean 0.976073 +trainer/policy/normal/log_std Std 0.318209 +trainer/policy/normal/log_std Max 1.67215 +trainer/policy/normal/log_std Min -1.20812 +eval/num steps total 729715 +eval/num paths total 732 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00532343 +eval/Actions Std 0.907708 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51497 +time/logging (s) 0.00372388 +time/sampling batch (s) 0.280497 +time/saving (s) 0.00351986 +time/training (s) 7.57555 +time/epoch (s) 10.3783 +time/total (s) 7587.87 +Epoch -269 +---------------------------------- --------------- +2022-05-10 15:17:24.971297 PDT | [1] Epoch -268 finished +---------------------------------- --------------- +epoch -268 +replay_buffer/size 999033 +trainer/num train calls 733000 +trainer/Policy Loss -19.0796 +trainer/Log Pis Mean 24.2422 +trainer/Log Pis Std 12.4828 +trainer/Log Pis Max 63.542 +trainer/Log Pis Min -7.04404 +trainer/policy/mean Mean -0.049731 +trainer/policy/mean Std 0.909556 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80647 +trainer/policy/normal/std Std 0.645193 +trainer/policy/normal/std Max 6.51255 +trainer/policy/normal/std Min 0.3164 +trainer/policy/normal/log_std Mean 0.995099 +trainer/policy/normal/log_std Std 0.303577 +trainer/policy/normal/log_std Max 1.87373 +trainer/policy/normal/log_std Min -1.15075 +eval/num steps total 730715 +eval/num paths total 733 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106534 +eval/Actions Std 0.908355 +eval/Actions Max 0.999992 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61905 +time/logging (s) 0.00379072 +time/sampling batch (s) 0.279776 +time/saving (s) 0.00347221 +time/training (s) 6.96252 +time/epoch (s) 9.86861 +time/total (s) 7597.74 +Epoch -268 +---------------------------------- --------------- +2022-05-10 15:17:34.694229 PDT | [1] Epoch -267 finished +---------------------------------- --------------- +epoch -267 +replay_buffer/size 999033 +trainer/num train calls 734000 +trainer/Policy Loss -19.2849 +trainer/Log Pis Mean 25.0639 +trainer/Log Pis Std 12.8802 +trainer/Log Pis Max 67.9945 +trainer/Log Pis Min -5.83994 +trainer/policy/mean Mean -0.0245782 +trainer/policy/mean Std 0.907865 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.822 +trainer/policy/normal/std Std 0.640788 +trainer/policy/normal/std Max 5.65356 +trainer/policy/normal/std Min 0.292336 +trainer/policy/normal/log_std Mean 1.0025 +trainer/policy/normal/log_std Std 0.293096 +trainer/policy/normal/log_std Max 1.73229 +trainer/policy/normal/log_std Min -1.22985 +eval/num steps total 731715 +eval/num paths total 734 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0241818 +eval/Actions Std 0.90185 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6003 +time/logging (s) 0.00440183 +time/sampling batch (s) 0.53017 +time/saving (s) 0.0042039 +time/training (s) 6.5606 +time/epoch (s) 9.69967 +time/total (s) 7607.45 +Epoch -267 +---------------------------------- --------------- +2022-05-10 15:17:44.437207 PDT | [1] Epoch -266 finished +---------------------------------- --------------- +epoch -266 +replay_buffer/size 999033 +trainer/num train calls 735000 +trainer/Policy Loss -19.7447 +trainer/Log Pis Mean 24.6992 +trainer/Log Pis Std 13.7059 +trainer/Log Pis Max 86.046 +trainer/Log Pis Min -6.28552 +trainer/policy/mean Mean -0.0871369 +trainer/policy/mean Std 0.907785 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86246 +trainer/policy/normal/std Std 0.680684 +trainer/policy/normal/std Max 6.70048 +trainer/policy/normal/std Min 0.282435 +trainer/policy/normal/log_std Mean 1.01134 +trainer/policy/normal/log_std Std 0.320679 +trainer/policy/normal/log_std Max 1.90218 +trainer/policy/normal/log_std Min -1.26431 +eval/num steps total 732715 +eval/num paths total 735 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227976 +eval/Actions Std 0.91583 +eval/Actions Max 0.999974 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48339 +time/logging (s) 0.00372585 +time/sampling batch (s) 0.278332 +time/saving (s) 0.00355489 +time/training (s) 6.94852 +time/epoch (s) 9.71753 +time/total (s) 7617.17 +Epoch -266 +---------------------------------- --------------- +2022-05-10 15:17:55.767997 PDT | [1] Epoch -265 finished +---------------------------------- --------------- +epoch -265 +replay_buffer/size 999033 +trainer/num train calls 736000 +trainer/Policy Loss -20.7498 +trainer/Log Pis Mean 24.9239 +trainer/Log Pis Std 13.1753 +trainer/Log Pis Max 75.6956 +trainer/Log Pis Min -5.40194 +trainer/policy/mean Mean -0.0235859 +trainer/policy/mean Std 0.910229 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83806 +trainer/policy/normal/std Std 0.66967 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.306619 +trainer/policy/normal/log_std Mean 1.00473 +trainer/policy/normal/log_std Std 0.311021 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.18215 +eval/num steps total 733715 +eval/num paths total 736 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.151806 +eval/Actions Std 0.86944 +eval/Actions Max 0.999994 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86149 +time/logging (s) 0.00367775 +time/sampling batch (s) 0.279155 +time/saving (s) 0.00340208 +time/training (s) 8.15883 +time/epoch (s) 11.3066 +time/total (s) 7628.48 +Epoch -265 +---------------------------------- --------------- +2022-05-10 15:18:06.124052 PDT | [1] Epoch -264 finished +---------------------------------- --------------- +epoch -264 +replay_buffer/size 999033 +trainer/num train calls 737000 +trainer/Policy Loss -18.1618 +trainer/Log Pis Mean 24.8902 +trainer/Log Pis Std 13.1448 +trainer/Log Pis Max 69.1202 +trainer/Log Pis Min -8.33386 +trainer/policy/mean Mean -0.016324 +trainer/policy/mean Std 0.906379 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83054 +trainer/policy/normal/std Std 0.666346 +trainer/policy/normal/std Max 5.95434 +trainer/policy/normal/std Min 0.262784 +trainer/policy/normal/log_std Mean 1.00195 +trainer/policy/normal/log_std Std 0.311194 +trainer/policy/normal/log_std Max 1.78412 +trainer/policy/normal/log_std Min -1.33642 +eval/num steps total 734715 +eval/num paths total 737 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.170304 +eval/Actions Std 0.868628 +eval/Actions Max 0.999989 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83139 +time/logging (s) 0.00454707 +time/sampling batch (s) 0.529804 +time/saving (s) 0.00436913 +time/training (s) 6.96283 +time/epoch (s) 10.3329 +time/total (s) 7638.81 +Epoch -264 +---------------------------------- --------------- +2022-05-10 15:18:16.204004 PDT | [1] Epoch -263 finished +---------------------------------- --------------- +epoch -263 +replay_buffer/size 999033 +trainer/num train calls 738000 +trainer/Policy Loss -19.1893 +trainer/Log Pis Mean 23.9919 +trainer/Log Pis Std 13.231 +trainer/Log Pis Max 79.9093 +trainer/Log Pis Min -10.6201 +trainer/policy/mean Mean -0.0266155 +trainer/policy/mean Std 0.905277 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.7623 +trainer/policy/normal/std Std 0.656366 +trainer/policy/normal/std Max 5.75905 +trainer/policy/normal/std Min 0.266072 +trainer/policy/normal/log_std Mean 0.97417 +trainer/policy/normal/log_std Std 0.329985 +trainer/policy/normal/log_std Max 1.75077 +trainer/policy/normal/log_std Min -1.32399 +eval/num steps total 735715 +eval/num paths total 738 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.224289 +eval/Actions Std 0.827015 +eval/Actions Max 0.999992 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77644 +time/logging (s) 0.00383978 +time/sampling batch (s) 0.540338 +time/saving (s) 0.00394993 +time/training (s) 6.7288 +time/epoch (s) 10.0534 +time/total (s) 7648.87 +Epoch -263 +---------------------------------- --------------- +2022-05-10 15:18:25.651761 PDT | [1] Epoch -262 finished +---------------------------------- --------------- +epoch -262 +replay_buffer/size 999033 +trainer/num train calls 739000 +trainer/Policy Loss -20.0534 +trainer/Log Pis Mean 25.9484 +trainer/Log Pis Std 13.5132 +trainer/Log Pis Max 67.4451 +trainer/Log Pis Min -7.10963 +trainer/policy/mean Mean -0.0369125 +trainer/policy/mean Std 0.906471 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83434 +trainer/policy/normal/std Std 0.66513 +trainer/policy/normal/std Max 6.33494 +trainer/policy/normal/std Min 0.22092 +trainer/policy/normal/log_std Mean 1.00192 +trainer/policy/normal/log_std Std 0.319585 +trainer/policy/normal/log_std Max 1.84608 +trainer/policy/normal/log_std Min -1.50995 +eval/num steps total 736715 +eval/num paths total 739 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0859951 +eval/Actions Std 0.929965 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49783 +time/logging (s) 0.00381559 +time/sampling batch (s) 0.291512 +time/saving (s) 0.00347305 +time/training (s) 6.62569 +time/epoch (s) 9.42232 +time/total (s) 7658.3 +Epoch -262 +---------------------------------- --------------- +2022-05-10 15:18:36.337749 PDT | [1] Epoch -261 finished +---------------------------------- --------------- +epoch -261 +replay_buffer/size 999033 +trainer/num train calls 740000 +trainer/Policy Loss -20.507 +trainer/Log Pis Mean 24.0987 +trainer/Log Pis Std 12.7114 +trainer/Log Pis Max 62.1162 +trainer/Log Pis Min -6.15138 +trainer/policy/mean Mean -0.0394086 +trainer/policy/mean Std 0.91256 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.7356 +trainer/policy/normal/std Std 0.627682 +trainer/policy/normal/std Max 5.12002 +trainer/policy/normal/std Min 0.224544 +trainer/policy/normal/log_std Mean 0.966807 +trainer/policy/normal/log_std Std 0.32095 +trainer/policy/normal/log_std Max 1.63316 +trainer/policy/normal/log_std Min -1.49368 +eval/num steps total 737715 +eval/num paths total 740 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.270082 +eval/Actions Std 0.917606 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54572 +time/logging (s) 0.00423183 +time/sampling batch (s) 0.278236 +time/saving (s) 0.00415659 +time/training (s) 7.82996 +time/epoch (s) 10.6623 +time/total (s) 7668.96 +Epoch -261 +---------------------------------- --------------- +2022-05-10 15:18:46.367015 PDT | [1] Epoch -260 finished +---------------------------------- --------------- +epoch -260 +replay_buffer/size 999033 +trainer/num train calls 741000 +trainer/Policy Loss -19.8534 +trainer/Log Pis Mean 24.5731 +trainer/Log Pis Std 12.8328 +trainer/Log Pis Max 74.9007 +trainer/Log Pis Min -7.58622 +trainer/policy/mean Mean -0.0430093 +trainer/policy/mean Std 0.906605 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74676 +trainer/policy/normal/std Std 0.632273 +trainer/policy/normal/std Max 5.39261 +trainer/policy/normal/std Min 0.282617 +trainer/policy/normal/log_std Mean 0.971943 +trainer/policy/normal/log_std Std 0.313702 +trainer/policy/normal/log_std Max 1.68503 +trainer/policy/normal/log_std Min -1.26366 +eval/num steps total 738715 +eval/num paths total 741 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0734634 +eval/Actions Std 0.900509 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58873 +time/logging (s) 0.00371414 +time/sampling batch (s) 0.27568 +time/saving (s) 0.0034382 +time/training (s) 7.13277 +time/epoch (s) 10.0043 +time/total (s) 7678.97 +Epoch -260 +---------------------------------- --------------- +2022-05-10 15:18:56.443302 PDT | [1] Epoch -259 finished +---------------------------------- --------------- +epoch -259 +replay_buffer/size 999033 +trainer/num train calls 742000 +trainer/Policy Loss -19.6997 +trainer/Log Pis Mean 25.217 +trainer/Log Pis Std 13.2894 +trainer/Log Pis Max 62.2962 +trainer/Log Pis Min -2.91322 +trainer/policy/mean Mean -0.0451048 +trainer/policy/mean Std 0.90725 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79226 +trainer/policy/normal/std Std 0.670251 +trainer/policy/normal/std Max 5.90413 +trainer/policy/normal/std Min 0.323505 +trainer/policy/normal/log_std Mean 0.985053 +trainer/policy/normal/log_std Std 0.326958 +trainer/policy/normal/log_std Max 1.77565 +trainer/policy/normal/log_std Min -1.12854 +eval/num steps total 739715 +eval/num paths total 742 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.274959 +eval/Actions Std 0.797303 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74877 +time/logging (s) 0.00369929 +time/sampling batch (s) 0.276203 +time/saving (s) 0.00349312 +time/training (s) 7.02039 +time/epoch (s) 10.0526 +time/total (s) 7689.02 +Epoch -259 +---------------------------------- --------------- +2022-05-10 15:19:06.685878 PDT | [1] Epoch -258 finished +---------------------------------- --------------- +epoch -258 +replay_buffer/size 999033 +trainer/num train calls 743000 +trainer/Policy Loss -20.3635 +trainer/Log Pis Mean 24.7756 +trainer/Log Pis Std 12.7 +trainer/Log Pis Max 64.8273 +trainer/Log Pis Min -5.25618 +trainer/policy/mean Mean -0.0439458 +trainer/policy/mean Std 0.905107 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81835 +trainer/policy/normal/std Std 0.659833 +trainer/policy/normal/std Max 6.65497 +trainer/policy/normal/std Min 0.20853 +trainer/policy/normal/log_std Mean 0.996414 +trainer/policy/normal/log_std Std 0.319333 +trainer/policy/normal/log_std Max 1.89536 +trainer/policy/normal/log_std Min -1.56767 +eval/num steps total 740715 +eval/num paths total 743 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0684079 +eval/Actions Std 0.91586 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58112 +time/logging (s) 0.004176 +time/sampling batch (s) 0.276505 +time/saving (s) 0.00398028 +time/training (s) 7.35343 +time/epoch (s) 10.2192 +time/total (s) 7699.25 +Epoch -258 +---------------------------------- --------------- +2022-05-10 15:19:16.987837 PDT | [1] Epoch -257 finished +---------------------------------- --------------- +epoch -257 +replay_buffer/size 999033 +trainer/num train calls 744000 +trainer/Policy Loss -19.8722 +trainer/Log Pis Mean 24.4796 +trainer/Log Pis Std 12.9713 +trainer/Log Pis Max 72.0369 +trainer/Log Pis Min -9.69114 +trainer/policy/mean Mean -0.0391117 +trainer/policy/mean Std 0.905361 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.74778 +trainer/policy/normal/std Std 0.640841 +trainer/policy/normal/std Max 6.21588 +trainer/policy/normal/std Min 0.239673 +trainer/policy/normal/log_std Mean 0.970368 +trainer/policy/normal/log_std Std 0.324677 +trainer/policy/normal/log_std Max 1.82711 +trainer/policy/normal/log_std Min -1.42848 +eval/num steps total 741715 +eval/num paths total 744 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0502776 +eval/Actions Std 0.909242 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39101 +time/logging (s) 0.00391173 +time/sampling batch (s) 0.277099 +time/saving (s) 0.00356145 +time/training (s) 7.60183 +time/epoch (s) 10.2774 +time/total (s) 7709.53 +Epoch -257 +---------------------------------- --------------- +2022-05-10 15:19:27.955729 PDT | [1] Epoch -256 finished +---------------------------------- --------------- +epoch -256 +replay_buffer/size 999033 +trainer/num train calls 745000 +trainer/Policy Loss -19.1858 +trainer/Log Pis Mean 25.2238 +trainer/Log Pis Std 13.262 +trainer/Log Pis Max 76.1861 +trainer/Log Pis Min -4.55313 +trainer/policy/mean Mean -0.0217481 +trainer/policy/mean Std 0.909139 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83223 +trainer/policy/normal/std Std 0.671311 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.206553 +trainer/policy/normal/log_std Mean 0.999508 +trainer/policy/normal/log_std Std 0.327941 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.5772 +eval/num steps total 742715 +eval/num paths total 745 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0882688 +eval/Actions Std 0.820992 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64466 +time/logging (s) 0.00370194 +time/sampling batch (s) 0.277152 +time/saving (s) 0.00341284 +time/training (s) 8.01482 +time/epoch (s) 10.9438 +time/total (s) 7720.47 +Epoch -256 +---------------------------------- --------------- +2022-05-10 15:19:38.301558 PDT | [1] Epoch -255 finished +---------------------------------- --------------- +epoch -255 +replay_buffer/size 999033 +trainer/num train calls 746000 +trainer/Policy Loss -20.3628 +trainer/Log Pis Mean 25.0101 +trainer/Log Pis Std 13.2918 +trainer/Log Pis Max 74.9165 +trainer/Log Pis Min -6.37911 +trainer/policy/mean Mean -0.0547915 +trainer/policy/mean Std 0.903613 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82436 +trainer/policy/normal/std Std 0.668628 +trainer/policy/normal/std Max 6.04089 +trainer/policy/normal/std Min 0.225381 +trainer/policy/normal/log_std Mean 0.994987 +trainer/policy/normal/log_std Std 0.339675 +trainer/policy/normal/log_std Max 1.79855 +trainer/policy/normal/log_std Min -1.48996 +eval/num steps total 743715 +eval/num paths total 746 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0786362 +eval/Actions Std 0.879831 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72843 +time/logging (s) 0.0037764 +time/sampling batch (s) 0.276884 +time/saving (s) 0.00354578 +time/training (s) 7.30935 +time/epoch (s) 10.322 +time/total (s) 7730.8 +Epoch -255 +---------------------------------- --------------- +2022-05-10 15:19:48.313894 PDT | [1] Epoch -254 finished +---------------------------------- --------------- +epoch -254 +replay_buffer/size 999033 +trainer/num train calls 747000 +trainer/Policy Loss -20.3289 +trainer/Log Pis Mean 23.7267 +trainer/Log Pis Std 13.261 +trainer/Log Pis Max 67.6972 +trainer/Log Pis Min -6.81473 +trainer/policy/mean Mean -0.0251301 +trainer/policy/mean Std 0.901809 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.78071 +trainer/policy/normal/std Std 0.682468 +trainer/policy/normal/std Max 5.30848 +trainer/policy/normal/std Min 0.238432 +trainer/policy/normal/log_std Mean 0.975806 +trainer/policy/normal/log_std Std 0.353826 +trainer/policy/normal/log_std Max 1.6693 +trainer/policy/normal/log_std Min -1.43367 +eval/num steps total 744715 +eval/num paths total 747 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138007 +eval/Actions Std 0.849639 +eval/Actions Max 0.999978 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56562 +time/logging (s) 0.0038329 +time/sampling batch (s) 0.27826 +time/saving (s) 0.00342698 +time/training (s) 7.13726 +time/epoch (s) 9.9884 +time/total (s) 7740.79 +Epoch -254 +---------------------------------- --------------- +2022-05-10 15:19:58.398775 PDT | [1] Epoch -253 finished +---------------------------------- --------------- +epoch -253 +replay_buffer/size 999033 +trainer/num train calls 748000 +trainer/Policy Loss -20.2461 +trainer/Log Pis Mean 24.7672 +trainer/Log Pis Std 13.6949 +trainer/Log Pis Max 70.544 +trainer/Log Pis Min -6.34588 +trainer/policy/mean Mean -0.0233998 +trainer/policy/mean Std 0.909296 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.76954 +trainer/policy/normal/std Std 0.683585 +trainer/policy/normal/std Max 5.98803 +trainer/policy/normal/std Min 0.267684 +trainer/policy/normal/log_std Mean 0.97313 +trainer/policy/normal/log_std Std 0.344954 +trainer/policy/normal/log_std Max 1.78976 +trainer/policy/normal/log_std Min -1.31795 +eval/num steps total 745715 +eval/num paths total 748 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106698 +eval/Actions Std 0.913272 +eval/Actions Max 0.999983 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48972 +time/logging (s) 0.0041459 +time/sampling batch (s) 0.281225 +time/saving (s) 0.00399031 +time/training (s) 7.28178 +time/epoch (s) 10.0609 +time/total (s) 7750.85 +Epoch -253 +---------------------------------- --------------- +2022-05-10 15:20:08.560829 PDT | [1] Epoch -252 finished +---------------------------------- --------------- +epoch -252 +replay_buffer/size 999033 +trainer/num train calls 749000 +trainer/Policy Loss -18.9967 +trainer/Log Pis Mean 24.49 +trainer/Log Pis Std 13.5689 +trainer/Log Pis Max 78.7917 +trainer/Log Pis Min -7.74347 +trainer/policy/mean Mean -0.0356951 +trainer/policy/mean Std 0.909794 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86913 +trainer/policy/normal/std Std 0.657215 +trainer/policy/normal/std Max 7.06893 +trainer/policy/normal/std Min 0.29001 +trainer/policy/normal/log_std Mean 1.01671 +trainer/policy/normal/log_std Std 0.308116 +trainer/policy/normal/log_std Max 1.95571 +trainer/policy/normal/log_std Min -1.23784 +eval/num steps total 746715 +eval/num paths total 749 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.091839 +eval/Actions Std 0.915252 +eval/Actions Max 0.999986 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57676 +time/logging (s) 0.00377302 +time/sampling batch (s) 0.276034 +time/saving (s) 0.00358986 +time/training (s) 7.27705 +time/epoch (s) 10.1372 +time/total (s) 7760.99 +Epoch -252 +---------------------------------- --------------- +2022-05-10 15:20:17.627605 PDT | [1] Epoch -251 finished +---------------------------------- --------------- +epoch -251 +replay_buffer/size 999033 +trainer/num train calls 750000 +trainer/Policy Loss -19.2923 +trainer/Log Pis Mean 24.1297 +trainer/Log Pis Std 12.2927 +trainer/Log Pis Max 67.3586 +trainer/Log Pis Min -7.81364 +trainer/policy/mean Mean -0.0427987 +trainer/policy/mean Std 0.908939 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.81043 +trainer/policy/normal/std Std 0.661558 +trainer/policy/normal/std Max 5.67684 +trainer/policy/normal/std Min 0.319012 +trainer/policy/normal/log_std Mean 0.99316 +trainer/policy/normal/log_std Std 0.320854 +trainer/policy/normal/log_std Max 1.7364 +trainer/policy/normal/log_std Min -1.14253 +eval/num steps total 747715 +eval/num paths total 750 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0678431 +eval/Actions Std 0.914437 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28692 +time/logging (s) 0.00374868 +time/sampling batch (s) 0.275707 +time/saving (s) 0.00349584 +time/training (s) 6.47287 +time/epoch (s) 9.04274 +time/total (s) 7770.04 +Epoch -251 +---------------------------------- --------------- +2022-05-10 15:20:27.663244 PDT | [1] Epoch -250 finished +---------------------------------- --------------- +epoch -250 +replay_buffer/size 999033 +trainer/num train calls 751000 +trainer/Policy Loss -19.6073 +trainer/Log Pis Mean 24.4907 +trainer/Log Pis Std 13.4398 +trainer/Log Pis Max 62.7335 +trainer/Log Pis Min -5.08124 +trainer/policy/mean Mean -0.03425 +trainer/policy/mean Std 0.908519 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83347 +trainer/policy/normal/std Std 0.697247 +trainer/policy/normal/std Max 7.07185 +trainer/policy/normal/std Min 0.309056 +trainer/policy/normal/log_std Mean 0.996988 +trainer/policy/normal/log_std Std 0.339332 +trainer/policy/normal/log_std Max 1.95612 +trainer/policy/normal/log_std Min -1.17423 +eval/num steps total 748715 +eval/num paths total 751 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0645987 +eval/Actions Std 0.913534 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46643 +time/logging (s) 0.00382063 +time/sampling batch (s) 0.525359 +time/saving (s) 0.00343956 +time/training (s) 7.01288 +time/epoch (s) 10.0119 +time/total (s) 7780.06 +Epoch -250 +---------------------------------- --------------- +2022-05-10 15:20:37.307941 PDT | [1] Epoch -249 finished +---------------------------------- --------------- +epoch -249 +replay_buffer/size 999033 +trainer/num train calls 752000 +trainer/Policy Loss -19.4562 +trainer/Log Pis Mean 24.7076 +trainer/Log Pis Std 12.4621 +trainer/Log Pis Max 66.8722 +trainer/Log Pis Min -6.74214 +trainer/policy/mean Mean -0.028298 +trainer/policy/mean Std 0.904624 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78628 +trainer/policy/normal/std Std 0.704928 +trainer/policy/normal/std Max 5.47856 +trainer/policy/normal/std Min 0.233561 +trainer/policy/normal/log_std Mean 0.976008 +trainer/policy/normal/log_std Std 0.35804 +trainer/policy/normal/log_std Max 1.70084 +trainer/policy/normal/log_std Min -1.45431 +eval/num steps total 749715 +eval/num paths total 752 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0781883 +eval/Actions Std 0.918381 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68668 +time/logging (s) 0.00413638 +time/sampling batch (s) 0.284567 +time/saving (s) 0.0040112 +time/training (s) 6.64112 +time/epoch (s) 9.62051 +time/total (s) 7789.68 +Epoch -249 +---------------------------------- --------------- +2022-05-10 15:20:46.680987 PDT | [1] Epoch -248 finished +---------------------------------- --------------- +epoch -248 +replay_buffer/size 999033 +trainer/num train calls 753000 +trainer/Policy Loss -18.8288 +trainer/Log Pis Mean 25.5033 +trainer/Log Pis Std 13.7909 +trainer/Log Pis Max 66.6394 +trainer/Log Pis Min -9.58351 +trainer/policy/mean Mean -0.0122055 +trainer/policy/mean Std 0.905342 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.853 +trainer/policy/normal/std Std 0.697659 +trainer/policy/normal/std Max 6.02687 +trainer/policy/normal/std Min 0.216047 +trainer/policy/normal/log_std Mean 1.0047 +trainer/policy/normal/log_std Std 0.335919 +trainer/policy/normal/log_std Max 1.79623 +trainer/policy/normal/log_std Min -1.53226 +eval/num steps total 750715 +eval/num paths total 753 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0200859 +eval/Actions Std 0.911329 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60799 +time/logging (s) 0.00409262 +time/sampling batch (s) 0.283616 +time/saving (s) 0.00356888 +time/training (s) 6.44892 +time/epoch (s) 9.34818 +time/total (s) 7799.03 +Epoch -248 +---------------------------------- --------------- +2022-05-10 15:20:57.940525 PDT | [1] Epoch -247 finished +---------------------------------- --------------- +epoch -247 +replay_buffer/size 999033 +trainer/num train calls 754000 +trainer/Policy Loss -19.6208 +trainer/Log Pis Mean 24.9417 +trainer/Log Pis Std 14.4001 +trainer/Log Pis Max 67.8529 +trainer/Log Pis Min -10.3262 +trainer/policy/mean Mean -0.0166523 +trainer/policy/mean Std 0.907399 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87269 +trainer/policy/normal/std Std 0.69431 +trainer/policy/normal/std Max 5.47591 +trainer/policy/normal/std Min 0.291732 +trainer/policy/normal/log_std Mean 1.01084 +trainer/policy/normal/log_std Std 0.342393 +trainer/policy/normal/log_std Max 1.70036 +trainer/policy/normal/log_std Min -1.23192 +eval/num steps total 751715 +eval/num paths total 754 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.019961 +eval/Actions Std 0.776118 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69742 +time/logging (s) 0.00421717 +time/sampling batch (s) 0.284388 +time/saving (s) 0.00389968 +time/training (s) 8.24407 +time/epoch (s) 11.234 +time/total (s) 7810.27 +Epoch -247 +---------------------------------- --------------- +2022-05-10 15:21:07.506200 PDT | [1] Epoch -246 finished +---------------------------------- --------------- +epoch -246 +replay_buffer/size 999033 +trainer/num train calls 755000 +trainer/Policy Loss -20.4676 +trainer/Log Pis Mean 25.0072 +trainer/Log Pis Std 13.4006 +trainer/Log Pis Max 69.3173 +trainer/Log Pis Min -2.55676 +trainer/policy/mean Mean -0.0324405 +trainer/policy/mean Std 0.903628 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81262 +trainer/policy/normal/std Std 0.687887 +trainer/policy/normal/std Max 5.59641 +trainer/policy/normal/std Min 0.214276 +trainer/policy/normal/log_std Mean 0.989494 +trainer/policy/normal/log_std Std 0.340752 +trainer/policy/normal/log_std Max 1.72212 +trainer/policy/normal/log_std Min -1.54049 +eval/num steps total 752715 +eval/num paths total 755 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129507 +eval/Actions Std 0.886778 +eval/Actions Max 0.999992 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72361 +time/logging (s) 0.00370645 +time/sampling batch (s) 0.28288 +time/saving (s) 0.00351598 +time/training (s) 6.52628 +time/epoch (s) 9.54 +time/total (s) 7819.81 +Epoch -246 +---------------------------------- --------------- +2022-05-10 15:21:18.057056 PDT | [1] Epoch -245 finished +---------------------------------- --------------- +epoch -245 +replay_buffer/size 999033 +trainer/num train calls 756000 +trainer/Policy Loss -17.2164 +trainer/Log Pis Mean 23.9011 +trainer/Log Pis Std 12.6527 +trainer/Log Pis Max 60.963 +trainer/Log Pis Min -6.39867 +trainer/policy/mean Mean -0.0214694 +trainer/policy/mean Std 0.904664 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83246 +trainer/policy/normal/std Std 0.687334 +trainer/policy/normal/std Max 6.43912 +trainer/policy/normal/std Min 0.228046 +trainer/policy/normal/log_std Mean 0.998364 +trainer/policy/normal/log_std Std 0.331655 +trainer/policy/normal/log_std Max 1.86239 +trainer/policy/normal/log_std Min -1.47821 +eval/num steps total 753715 +eval/num paths total 756 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.107276 +eval/Actions Std 0.826517 +eval/Actions Max 0.999982 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67206 +time/logging (s) 0.0037239 +time/sampling batch (s) 0.777694 +time/saving (s) 0.00342812 +time/training (s) 7.06988 +time/epoch (s) 10.5268 +time/total (s) 7830.34 +Epoch -245 +---------------------------------- --------------- +2022-05-10 15:21:28.644127 PDT | [1] Epoch -244 finished +---------------------------------- --------------- +epoch -244 +replay_buffer/size 999033 +trainer/num train calls 757000 +trainer/Policy Loss -19.6133 +trainer/Log Pis Mean 23.9045 +trainer/Log Pis Std 13.6718 +trainer/Log Pis Max 68.7949 +trainer/Log Pis Min -6.47608 +trainer/policy/mean Mean -0.0444932 +trainer/policy/mean Std 0.904831 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82688 +trainer/policy/normal/std Std 0.659625 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.307369 +trainer/policy/normal/log_std Mean 1.00035 +trainer/policy/normal/log_std Std 0.314024 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.17971 +eval/num steps total 754715 +eval/num paths total 757 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.18444 +eval/Actions Std 0.940078 +eval/Actions Max 0.999984 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62865 +time/logging (s) 0.00375247 +time/sampling batch (s) 0.778904 +time/saving (s) 0.00347231 +time/training (s) 7.1482 +time/epoch (s) 10.563 +time/total (s) 7840.91 +Epoch -244 +---------------------------------- --------------- +2022-05-10 15:21:38.653168 PDT | [1] Epoch -243 finished +---------------------------------- --------------- +epoch -243 +replay_buffer/size 999033 +trainer/num train calls 758000 +trainer/Policy Loss -20.2026 +trainer/Log Pis Mean 25.0459 +trainer/Log Pis Std 13.628 +trainer/Log Pis Max 73.6142 +trainer/Log Pis Min -7.16066 +trainer/policy/mean Mean -0.0448968 +trainer/policy/mean Std 0.905597 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82242 +trainer/policy/normal/std Std 0.683151 +trainer/policy/normal/std Max 6.24484 +trainer/policy/normal/std Min 0.226766 +trainer/policy/normal/log_std Mean 0.994421 +trainer/policy/normal/log_std Std 0.3352 +trainer/policy/normal/log_std Max 1.83176 +trainer/policy/normal/log_std Min -1.48384 +eval/num steps total 755715 +eval/num paths total 758 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.14138 +eval/Actions Std 0.906593 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55408 +time/logging (s) 0.0041158 +time/sampling batch (s) 0.531736 +time/saving (s) 0.00380716 +time/training (s) 6.89118 +time/epoch (s) 9.98492 +time/total (s) 7850.89 +Epoch -243 +---------------------------------- --------------- +2022-05-10 15:21:48.733686 PDT | [1] Epoch -242 finished +---------------------------------- --------------- +epoch -242 +replay_buffer/size 999033 +trainer/num train calls 759000 +trainer/Policy Loss -19.8421 +trainer/Log Pis Mean 23.6783 +trainer/Log Pis Std 13.6458 +trainer/Log Pis Max 73.7211 +trainer/Log Pis Min -7.35197 +trainer/policy/mean Mean -0.0282002 +trainer/policy/mean Std 0.90513 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82524 +trainer/policy/normal/std Std 0.707559 +trainer/policy/normal/std Max 6.40048 +trainer/policy/normal/std Min 0.235741 +trainer/policy/normal/log_std Mean 0.98961 +trainer/policy/normal/log_std Std 0.362477 +trainer/policy/normal/log_std Max 1.85637 +trainer/policy/normal/log_std Min -1.44502 +eval/num steps total 756715 +eval/num paths total 759 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0786215 +eval/Actions Std 0.913217 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54526 +time/logging (s) 0.00387542 +time/sampling batch (s) 1.02757 +time/saving (s) 0.00376166 +time/training (s) 6.47534 +time/epoch (s) 10.0558 +time/total (s) 7860.95 +Epoch -242 +---------------------------------- --------------- +2022-05-10 15:21:58.174990 PDT | [1] Epoch -241 finished +---------------------------------- --------------- +epoch -241 +replay_buffer/size 999033 +trainer/num train calls 760000 +trainer/Policy Loss -20.3684 +trainer/Log Pis Mean 24.4327 +trainer/Log Pis Std 12.824 +trainer/Log Pis Max 70.687 +trainer/Log Pis Min -10.31 +trainer/policy/mean Mean -0.0477938 +trainer/policy/mean Std 0.908766 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83282 +trainer/policy/normal/std Std 0.671953 +trainer/policy/normal/std Max 6.06033 +trainer/policy/normal/std Min 0.298778 +trainer/policy/normal/log_std Mean 1.0024 +trainer/policy/normal/log_std Std 0.311313 +trainer/policy/normal/log_std Max 1.80176 +trainer/policy/normal/log_std Min -1.20805 +eval/num steps total 757715 +eval/num paths total 760 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.200549 +eval/Actions Std 0.826207 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73253 +time/logging (s) 0.00376846 +time/sampling batch (s) 0.297988 +time/saving (s) 0.0035953 +time/training (s) 6.37876 +time/epoch (s) 9.41664 +time/total (s) 7870.37 +Epoch -241 +---------------------------------- --------------- +2022-05-10 15:22:08.124085 PDT | [1] Epoch -240 finished +---------------------------------- --------------- +epoch -240 +replay_buffer/size 999033 +trainer/num train calls 761000 +trainer/Policy Loss -19.561 +trainer/Log Pis Mean 25.5926 +trainer/Log Pis Std 13.1924 +trainer/Log Pis Max 70.5275 +trainer/Log Pis Min -7.20275 +trainer/policy/mean Mean -0.054956 +trainer/policy/mean Std 0.907683 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86716 +trainer/policy/normal/std Std 0.684847 +trainer/policy/normal/std Max 6.47379 +trainer/policy/normal/std Min 0.28208 +trainer/policy/normal/log_std Mean 1.01151 +trainer/policy/normal/log_std Std 0.329339 +trainer/policy/normal/log_std Max 1.86776 +trainer/policy/normal/log_std Min -1.26557 +eval/num steps total 758715 +eval/num paths total 761 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0868165 +eval/Actions Std 0.914895 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60538 +time/logging (s) 0.00370804 +time/sampling batch (s) 0.528234 +time/saving (s) 0.00348089 +time/training (s) 6.78427 +time/epoch (s) 9.92507 +time/total (s) 7880.3 +Epoch -240 +---------------------------------- --------------- +2022-05-10 15:22:18.441209 PDT | [1] Epoch -239 finished +---------------------------------- --------------- +epoch -239 +replay_buffer/size 999033 +trainer/num train calls 762000 +trainer/Policy Loss -19.6406 +trainer/Log Pis Mean 24.7046 +trainer/Log Pis Std 12.7784 +trainer/Log Pis Max 62.0479 +trainer/Log Pis Min -4.88511 +trainer/policy/mean Mean -0.0209788 +trainer/policy/mean Std 0.909532 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83161 +trainer/policy/normal/std Std 0.693176 +trainer/policy/normal/std Max 5.68759 +trainer/policy/normal/std Min 0.315079 +trainer/policy/normal/log_std Mean 0.995449 +trainer/policy/normal/log_std Std 0.344408 +trainer/policy/normal/log_std Max 1.73829 +trainer/policy/normal/log_std Min -1.15493 +eval/num steps total 759715 +eval/num paths total 762 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0753444 +eval/Actions Std 0.912627 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72381 +time/logging (s) 0.00442347 +time/sampling batch (s) 0.279425 +time/saving (s) 0.00413658 +time/training (s) 7.28196 +time/epoch (s) 10.2938 +time/total (s) 7890.6 +Epoch -239 +---------------------------------- --------------- +2022-05-10 15:22:28.261858 PDT | [1] Epoch -238 finished +---------------------------------- --------------- +epoch -238 +replay_buffer/size 999033 +trainer/num train calls 763000 +trainer/Policy Loss -19.9705 +trainer/Log Pis Mean 23.3472 +trainer/Log Pis Std 13.0285 +trainer/Log Pis Max 74.4452 +trainer/Log Pis Min -5.75957 +trainer/policy/mean Mean -0.0171147 +trainer/policy/mean Std 0.905954 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78657 +trainer/policy/normal/std Std 0.657759 +trainer/policy/normal/std Max 5.29194 +trainer/policy/normal/std Min 0.248547 +trainer/policy/normal/log_std Mean 0.983822 +trainer/policy/normal/log_std Std 0.325784 +trainer/policy/normal/log_std Max 1.66619 +trainer/policy/normal/log_std Min -1.39212 +eval/num steps total 760715 +eval/num paths total 763 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0569401 +eval/Actions Std 0.90595 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75601 +time/logging (s) 0.00374197 +time/sampling batch (s) 0.275546 +time/saving (s) 0.00342685 +time/training (s) 6.75688 +time/epoch (s) 9.79561 +time/total (s) 7900.4 +Epoch -238 +---------------------------------- --------------- +2022-05-10 15:22:38.311719 PDT | [1] Epoch -237 finished +---------------------------------- --------------- +epoch -237 +replay_buffer/size 999033 +trainer/num train calls 764000 +trainer/Policy Loss -20.1546 +trainer/Log Pis Mean 23.6808 +trainer/Log Pis Std 13.8425 +trainer/Log Pis Max 60.5455 +trainer/Log Pis Min -11.7576 +trainer/policy/mean Mean -0.0426307 +trainer/policy/mean Std 0.906707 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78763 +trainer/policy/normal/std Std 0.683986 +trainer/policy/normal/std Max 6.49601 +trainer/policy/normal/std Min 0.251899 +trainer/policy/normal/log_std Mean 0.981269 +trainer/policy/normal/log_std Std 0.335802 +trainer/policy/normal/log_std Max 1.87119 +trainer/policy/normal/log_std Min -1.37873 +eval/num steps total 761715 +eval/num paths total 764 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0104695 +eval/Actions Std 0.906929 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49582 +time/logging (s) 0.00413984 +time/sampling batch (s) 0.276187 +time/saving (s) 0.00398725 +time/training (s) 7.24627 +time/epoch (s) 10.0264 +time/total (s) 7910.43 +Epoch -237 +---------------------------------- --------------- +2022-05-10 15:22:49.592540 PDT | [1] Epoch -236 finished +---------------------------------- --------------- +epoch -236 +replay_buffer/size 999033 +trainer/num train calls 765000 +trainer/Policy Loss -19.6221 +trainer/Log Pis Mean 24.942 +trainer/Log Pis Std 13.1078 +trainer/Log Pis Max 69.6411 +trainer/Log Pis Min -9.39755 +trainer/policy/mean Mean -0.0437205 +trainer/policy/mean Std 0.909605 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86098 +trainer/policy/normal/std Std 0.701604 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.26666 +trainer/policy/normal/log_std Mean 1.00736 +trainer/policy/normal/log_std Std 0.336744 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.32178 +eval/num steps total 762715 +eval/num paths total 765 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.523301 +eval/Actions Std 0.752351 +eval/Actions Max 0.999998 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66391 +time/logging (s) 0.00387173 +time/sampling batch (s) 0.781484 +time/saving (s) 0.00358454 +time/training (s) 7.80298 +time/epoch (s) 11.2558 +time/total (s) 7921.69 +Epoch -236 +---------------------------------- --------------- +2022-05-10 15:22:59.005269 PDT | [1] Epoch -235 finished +---------------------------------- --------------- +epoch -235 +replay_buffer/size 999033 +trainer/num train calls 766000 +trainer/Policy Loss -19.9948 +trainer/Log Pis Mean 24.1785 +trainer/Log Pis Std 13.2882 +trainer/Log Pis Max 65.398 +trainer/Log Pis Min -7.26317 +trainer/policy/mean Mean -0.0442124 +trainer/policy/mean Std 0.902775 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78605 +trainer/policy/normal/std Std 0.687604 +trainer/policy/normal/std Max 6.28537 +trainer/policy/normal/std Min 0.226862 +trainer/policy/normal/log_std Mean 0.980698 +trainer/policy/normal/log_std Std 0.335982 +trainer/policy/normal/log_std Max 1.83823 +trainer/policy/normal/log_std Min -1.48341 +eval/num steps total 763715 +eval/num paths total 766 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102421 +eval/Actions Std 0.905727 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53385 +time/logging (s) 0.00376216 +time/sampling batch (s) 0.280902 +time/saving (s) 0.00345183 +time/training (s) 6.56639 +time/epoch (s) 9.38836 +time/total (s) 7931.08 +Epoch -235 +---------------------------------- --------------- +2022-05-10 15:23:08.724764 PDT | [1] Epoch -234 finished +---------------------------------- --------------- +epoch -234 +replay_buffer/size 999033 +trainer/num train calls 767000 +trainer/Policy Loss -21.116 +trainer/Log Pis Mean 23.5983 +trainer/Log Pis Std 13.9657 +trainer/Log Pis Max 67.6481 +trainer/Log Pis Min -8.52841 +trainer/policy/mean Mean -0.046623 +trainer/policy/mean Std 0.906145 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75545 +trainer/policy/normal/std Std 0.656227 +trainer/policy/normal/std Max 5.29884 +trainer/policy/normal/std Min 0.287425 +trainer/policy/normal/log_std Mean 0.971839 +trainer/policy/normal/log_std Std 0.328534 +trainer/policy/normal/log_std Max 1.66749 +trainer/policy/normal/log_std Min -1.24679 +eval/num steps total 764715 +eval/num paths total 767 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.043336 +eval/Actions Std 0.902411 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57687 +time/logging (s) 0.00409251 +time/sampling batch (s) 0.277883 +time/saving (s) 0.00386624 +time/training (s) 6.83297 +time/epoch (s) 9.69568 +time/total (s) 7940.78 +Epoch -234 +---------------------------------- --------------- +2022-05-10 15:23:19.462784 PDT | [1] Epoch -233 finished +---------------------------------- --------------- +epoch -233 +replay_buffer/size 999033 +trainer/num train calls 768000 +trainer/Policy Loss -19.0186 +trainer/Log Pis Mean 24.8685 +trainer/Log Pis Std 13.6723 +trainer/Log Pis Max 66.0922 +trainer/Log Pis Min -6.20862 +trainer/policy/mean Mean -0.0574834 +trainer/policy/mean Std 0.905425 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.82041 +trainer/policy/normal/std Std 0.673746 +trainer/policy/normal/std Max 6.20836 +trainer/policy/normal/std Min 0.251817 +trainer/policy/normal/log_std Mean 0.99451 +trainer/policy/normal/log_std Std 0.332502 +trainer/policy/normal/log_std Max 1.8259 +trainer/policy/normal/log_std Min -1.37905 +eval/num steps total 765715 +eval/num paths total 768 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.307819 +eval/Actions Std 0.837206 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66508 +time/logging (s) 0.00387034 +time/sampling batch (s) 0.284624 +time/saving (s) 0.00372683 +time/training (s) 7.7556 +time/epoch (s) 10.7129 +time/total (s) 7951.49 +Epoch -233 +---------------------------------- --------------- +2022-05-10 15:23:29.265424 PDT | [1] Epoch -232 finished +---------------------------------- --------------- +epoch -232 +replay_buffer/size 999033 +trainer/num train calls 769000 +trainer/Policy Loss -19.1009 +trainer/Log Pis Mean 25.3635 +trainer/Log Pis Std 14.2785 +trainer/Log Pis Max 65.0257 +trainer/Log Pis Min -4.52495 +trainer/policy/mean Mean -0.0548928 +trainer/policy/mean Std 0.907463 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88369 +trainer/policy/normal/std Std 0.692712 +trainer/policy/normal/std Max 5.73901 +trainer/policy/normal/std Min 0.264084 +trainer/policy/normal/log_std Mean 1.01706 +trainer/policy/normal/log_std Std 0.328837 +trainer/policy/normal/log_std Max 1.74729 +trainer/policy/normal/log_std Min -1.33149 +eval/num steps total 766715 +eval/num paths total 769 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0765769 +eval/Actions Std 0.906595 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75403 +time/logging (s) 0.00380215 +time/sampling batch (s) 0.279191 +time/saving (s) 0.00344211 +time/training (s) 6.73779 +time/epoch (s) 9.77826 +time/total (s) 7961.27 +Epoch -232 +---------------------------------- --------------- +2022-05-10 15:23:39.772690 PDT | [1] Epoch -231 finished +---------------------------------- --------------- +epoch -231 +replay_buffer/size 999033 +trainer/num train calls 770000 +trainer/Policy Loss -20.2769 +trainer/Log Pis Mean 24.1234 +trainer/Log Pis Std 13.5542 +trainer/Log Pis Max 74.1399 +trainer/Log Pis Min -7.12541 +trainer/policy/mean Mean -0.0391736 +trainer/policy/mean Std 0.90438 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81829 +trainer/policy/normal/std Std 0.687465 +trainer/policy/normal/std Max 7.1725 +trainer/policy/normal/std Min 0.266503 +trainer/policy/normal/log_std Mean 0.991831 +trainer/policy/normal/log_std Std 0.340394 +trainer/policy/normal/log_std Max 1.97025 +trainer/policy/normal/log_std Min -1.32237 +eval/num steps total 767715 +eval/num paths total 770 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0240447 +eval/Actions Std 0.90548 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57603 +time/logging (s) 0.00423841 +time/sampling batch (s) 0.308572 +time/saving (s) 0.00425294 +time/training (s) 7.58965 +time/epoch (s) 10.4828 +time/total (s) 7971.76 +Epoch -231 +---------------------------------- --------------- +2022-05-10 15:23:50.764545 PDT | [1] Epoch -230 finished +---------------------------------- --------------- +epoch -230 +replay_buffer/size 999033 +trainer/num train calls 771000 +trainer/Policy Loss -18.9376 +trainer/Log Pis Mean 24.9255 +trainer/Log Pis Std 13.6462 +trainer/Log Pis Max 67.0131 +trainer/Log Pis Min -9.00239 +trainer/policy/mean Mean -0.0194682 +trainer/policy/mean Std 0.905191 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80254 +trainer/policy/normal/std Std 0.664783 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.262149 +trainer/policy/normal/log_std Mean 0.989497 +trainer/policy/normal/log_std Std 0.325079 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.33884 +eval/num steps total 768715 +eval/num paths total 771 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0546293 +eval/Actions Std 0.912804 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73797 +time/logging (s) 0.00395259 +time/sampling batch (s) 0.587147 +time/saving (s) 0.00382857 +time/training (s) 7.63289 +time/epoch (s) 10.9658 +time/total (s) 7982.73 +Epoch -230 +---------------------------------- --------------- +2022-05-10 15:24:02.446614 PDT | [1] Epoch -229 finished +---------------------------------- --------------- +epoch -229 +replay_buffer/size 999033 +trainer/num train calls 772000 +trainer/Policy Loss -18.9754 +trainer/Log Pis Mean 23.2088 +trainer/Log Pis Std 13.239 +trainer/Log Pis Max 65.799 +trainer/Log Pis Min -3.30339 +trainer/policy/mean Mean -0.0466571 +trainer/policy/mean Std 0.901507 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.70877 +trainer/policy/normal/std Std 0.639362 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.344527 +trainer/policy/normal/log_std Mean 0.958167 +trainer/policy/normal/log_std Std 0.308618 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.06558 +eval/num steps total 769715 +eval/num paths total 772 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.013666 +eval/Actions Std 0.911873 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64055 +time/logging (s) 0.00368566 +time/sampling batch (s) 0.290456 +time/saving (s) 0.00340985 +time/training (s) 8.71896 +time/epoch (s) 11.6571 +time/total (s) 7994.39 +Epoch -229 +---------------------------------- --------------- +2022-05-10 15:24:12.367033 PDT | [1] Epoch -228 finished +---------------------------------- --------------- +epoch -228 +replay_buffer/size 999033 +trainer/num train calls 773000 +trainer/Policy Loss -19.7919 +trainer/Log Pis Mean 24.6599 +trainer/Log Pis Std 13.7941 +trainer/Log Pis Max 69.0914 +trainer/Log Pis Min -11.5569 +trainer/policy/mean Mean -0.0399393 +trainer/policy/mean Std 0.898891 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79002 +trainer/policy/normal/std Std 0.669403 +trainer/policy/normal/std Max 5.6332 +trainer/policy/normal/std Min 0.205853 +trainer/policy/normal/log_std Mean 0.981738 +trainer/policy/normal/log_std Std 0.342997 +trainer/policy/normal/log_std Max 1.72868 +trainer/policy/normal/log_std Min -1.58059 +eval/num steps total 770715 +eval/num paths total 773 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0116565 +eval/Actions Std 0.934006 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67751 +time/logging (s) 0.0036925 +time/sampling batch (s) 0.528471 +time/saving (s) 0.00344459 +time/training (s) 6.6833 +time/epoch (s) 9.89642 +time/total (s) 8004.29 +Epoch -228 +---------------------------------- --------------- +2022-05-10 15:24:22.061323 PDT | [1] Epoch -227 finished +---------------------------------- --------------- +epoch -227 +replay_buffer/size 999033 +trainer/num train calls 774000 +trainer/Policy Loss -20.0771 +trainer/Log Pis Mean 24.9348 +trainer/Log Pis Std 12.6955 +trainer/Log Pis Max 66.6053 +trainer/Log Pis Min -9.64024 +trainer/policy/mean Mean -0.0437739 +trainer/policy/mean Std 0.908709 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83059 +trainer/policy/normal/std Std 0.694235 +trainer/policy/normal/std Max 6.86222 +trainer/policy/normal/std Min 0.279218 +trainer/policy/normal/log_std Mean 0.996067 +trainer/policy/normal/log_std Std 0.339411 +trainer/policy/normal/log_std Max 1.92603 +trainer/policy/normal/log_std Min -1.27576 +eval/num steps total 771715 +eval/num paths total 774 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.335266 +eval/Actions Std 0.85744 +eval/Actions Max 0.999958 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69073 +time/logging (s) 0.00419074 +time/sampling batch (s) 0.276185 +time/saving (s) 0.00393081 +time/training (s) 6.69577 +time/epoch (s) 9.67081 +time/total (s) 8013.96 +Epoch -227 +---------------------------------- --------------- +2022-05-10 15:24:32.613056 PDT | [1] Epoch -226 finished +---------------------------------- --------------- +epoch -226 +replay_buffer/size 999033 +trainer/num train calls 775000 +trainer/Policy Loss -19.0198 +trainer/Log Pis Mean 23.8027 +trainer/Log Pis Std 13.0445 +trainer/Log Pis Max 64.827 +trainer/Log Pis Min -10.1005 +trainer/policy/mean Mean -0.0217525 +trainer/policy/mean Std 0.906153 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.74862 +trainer/policy/normal/std Std 0.66204 +trainer/policy/normal/std Max 5.7622 +trainer/policy/normal/std Min 0.240075 +trainer/policy/normal/log_std Mean 0.967675 +trainer/policy/normal/log_std Std 0.337665 +trainer/policy/normal/log_std Max 1.75132 +trainer/policy/normal/log_std Min -1.4268 +eval/num steps total 772715 +eval/num paths total 775 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0258614 +eval/Actions Std 0.967737 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59346 +time/logging (s) 0.00371315 +time/sampling batch (s) 0.277434 +time/saving (s) 0.00354512 +time/training (s) 7.64882 +time/epoch (s) 10.527 +time/total (s) 8024.49 +Epoch -226 +---------------------------------- --------------- +2022-05-10 15:24:42.533184 PDT | [1] Epoch -225 finished +---------------------------------- --------------- +epoch -225 +replay_buffer/size 999033 +trainer/num train calls 776000 +trainer/Policy Loss -19.7831 +trainer/Log Pis Mean 25.0628 +trainer/Log Pis Std 13.0225 +trainer/Log Pis Max 76.6114 +trainer/Log Pis Min -5.92017 +trainer/policy/mean Mean -0.0371231 +trainer/policy/mean Std 0.908838 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7957 +trainer/policy/normal/std Std 0.701587 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.304237 +trainer/policy/normal/log_std Mean 0.982334 +trainer/policy/normal/log_std Std 0.343115 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.18995 +eval/num steps total 773715 +eval/num paths total 776 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.397151 +eval/Actions Std 0.898089 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68891 +time/logging (s) 0.00366449 +time/sampling batch (s) 0.277119 +time/saving (s) 0.00343104 +time/training (s) 6.92284 +time/epoch (s) 9.89597 +time/total (s) 8034.39 +Epoch -225 +---------------------------------- --------------- +2022-05-10 15:24:53.413335 PDT | [1] Epoch -224 finished +---------------------------------- --------------- +epoch -224 +replay_buffer/size 999033 +trainer/num train calls 777000 +trainer/Policy Loss -19.7248 +trainer/Log Pis Mean 24.6269 +trainer/Log Pis Std 14.2417 +trainer/Log Pis Max 66.9203 +trainer/Log Pis Min -6.90517 +trainer/policy/mean Mean -0.0234267 +trainer/policy/mean Std 0.905517 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.82851 +trainer/policy/normal/std Std 0.708186 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.283663 +trainer/policy/normal/log_std Mean 0.992905 +trainer/policy/normal/log_std Std 0.349742 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.25997 +eval/num steps total 774715 +eval/num paths total 777 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00608534 +eval/Actions Std 0.911045 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58 +time/logging (s) 0.00364995 +time/sampling batch (s) 0.529336 +time/saving (s) 0.00344874 +time/training (s) 7.73945 +time/epoch (s) 10.8559 +time/total (s) 8045.25 +Epoch -224 +---------------------------------- --------------- +2022-05-10 15:25:04.650686 PDT | [1] Epoch -223 finished +---------------------------------- --------------- +epoch -223 +replay_buffer/size 999033 +trainer/num train calls 778000 +trainer/Policy Loss -18.99 +trainer/Log Pis Mean 24.2774 +trainer/Log Pis Std 13.4278 +trainer/Log Pis Max 63.5617 +trainer/Log Pis Min -7.11914 +trainer/policy/mean Mean -0.0245826 +trainer/policy/mean Std 0.908025 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77764 +trainer/policy/normal/std Std 0.65521 +trainer/policy/normal/std Max 5.35623 +trainer/policy/normal/std Min 0.286474 +trainer/policy/normal/log_std Mean 0.98032 +trainer/policy/normal/log_std Std 0.326347 +trainer/policy/normal/log_std Max 1.67826 +trainer/policy/normal/log_std Min -1.25011 +eval/num steps total 775715 +eval/num paths total 778 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0228478 +eval/Actions Std 0.962321 +eval/Actions Max 0.999996 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6168 +time/logging (s) 0.00410554 +time/sampling batch (s) 0.529856 +time/saving (s) 0.00396532 +time/training (s) 8.05858 +time/epoch (s) 11.2133 +time/total (s) 8056.47 +Epoch -223 +---------------------------------- --------------- +2022-05-10 15:25:15.693084 PDT | [1] Epoch -222 finished +---------------------------------- --------------- +epoch -222 +replay_buffer/size 999033 +trainer/num train calls 779000 +trainer/Policy Loss -21.1713 +trainer/Log Pis Mean 24.8415 +trainer/Log Pis Std 13.6504 +trainer/Log Pis Max 77.1569 +trainer/Log Pis Min -9.69964 +trainer/policy/mean Mean -0.0258562 +trainer/policy/mean Std 0.909321 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83682 +trainer/policy/normal/std Std 0.691325 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.312247 +trainer/policy/normal/log_std Mean 1.0001 +trainer/policy/normal/log_std Std 0.328893 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.16396 +eval/num steps total 776715 +eval/num paths total 779 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.236663 +eval/Actions Std 0.868501 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56772 +time/logging (s) 0.00384124 +time/sampling batch (s) 0.277781 +time/saving (s) 0.0036075 +time/training (s) 8.16457 +time/epoch (s) 11.0175 +time/total (s) 8067.49 +Epoch -222 +---------------------------------- --------------- +2022-05-10 15:25:26.820827 PDT | [1] Epoch -221 finished +---------------------------------- --------------- +epoch -221 +replay_buffer/size 999033 +trainer/num train calls 780000 +trainer/Policy Loss -20.0313 +trainer/Log Pis Mean 25.1612 +trainer/Log Pis Std 13.1181 +trainer/Log Pis Max 67.9897 +trainer/Log Pis Min -3.75724 +trainer/policy/mean Mean -0.0477253 +trainer/policy/mean Std 0.905223 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8349 +trainer/policy/normal/std Std 0.685783 +trainer/policy/normal/std Max 5.57366 +trainer/policy/normal/std Min 0.291749 +trainer/policy/normal/log_std Mean 0.998927 +trainer/policy/normal/log_std Std 0.33378 +trainer/policy/normal/log_std Max 1.71805 +trainer/policy/normal/log_std Min -1.23186 +eval/num steps total 777715 +eval/num paths total 780 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.197798 +eval/Actions Std 0.95508 +eval/Actions Max 0.999986 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75602 +time/logging (s) 0.00376685 +time/sampling batch (s) 0.277096 +time/saving (s) 0.00339392 +time/training (s) 8.06316 +time/epoch (s) 11.1034 +time/total (s) 8078.59 +Epoch -221 +---------------------------------- --------------- +2022-05-10 15:25:36.395154 PDT | [1] Epoch -220 finished +---------------------------------- --------------- +epoch -220 +replay_buffer/size 999033 +trainer/num train calls 781000 +trainer/Policy Loss -18.9491 +trainer/Log Pis Mean 24.4419 +trainer/Log Pis Std 13.3432 +trainer/Log Pis Max 64.4254 +trainer/Log Pis Min -10.1373 +trainer/policy/mean Mean -0.0468879 +trainer/policy/mean Std 0.910849 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76831 +trainer/policy/normal/std Std 0.619326 +trainer/policy/normal/std Max 6.54648 +trainer/policy/normal/std Min 0.267962 +trainer/policy/normal/log_std Mean 0.982599 +trainer/policy/normal/log_std Std 0.301318 +trainer/policy/normal/log_std Max 1.87893 +trainer/policy/normal/log_std Min -1.31691 +eval/num steps total 778715 +eval/num paths total 781 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.29735 +eval/Actions Std 0.8849 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58304 +time/logging (s) 0.00374967 +time/sampling batch (s) 0.276922 +time/saving (s) 0.00347902 +time/training (s) 6.68299 +time/epoch (s) 9.55018 +time/total (s) 8088.15 +Epoch -220 +---------------------------------- --------------- +2022-05-10 15:25:46.716133 PDT | [1] Epoch -219 finished +---------------------------------- --------------- +epoch -219 +replay_buffer/size 999033 +trainer/num train calls 782000 +trainer/Policy Loss -19.0868 +trainer/Log Pis Mean 23.8293 +trainer/Log Pis Std 12.8883 +trainer/Log Pis Max 65.8249 +trainer/Log Pis Min -5.03407 +trainer/policy/mean Mean -0.0276951 +trainer/policy/mean Std 0.905198 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.82726 +trainer/policy/normal/std Std 0.688485 +trainer/policy/normal/std Max 5.71757 +trainer/policy/normal/std Min 0.280628 +trainer/policy/normal/log_std Mean 0.995047 +trainer/policy/normal/log_std Std 0.339261 +trainer/policy/normal/log_std Max 1.74354 +trainer/policy/normal/log_std Min -1.27073 +eval/num steps total 779715 +eval/num paths total 782 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0643277 +eval/Actions Std 0.896251 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64548 +time/logging (s) 0.0037374 +time/sampling batch (s) 0.526419 +time/saving (s) 0.00340991 +time/training (s) 7.11783 +time/epoch (s) 10.2969 +time/total (s) 8098.45 +Epoch -219 +---------------------------------- --------------- +2022-05-10 15:25:55.911673 PDT | [1] Epoch -218 finished +---------------------------------- --------------- +epoch -218 +replay_buffer/size 999033 +trainer/num train calls 783000 +trainer/Policy Loss -17.9173 +trainer/Log Pis Mean 24.2346 +trainer/Log Pis Std 12.9438 +trainer/Log Pis Max 65.8006 +trainer/Log Pis Min -7.81781 +trainer/policy/mean Mean -0.0479635 +trainer/policy/mean Std 0.901842 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79907 +trainer/policy/normal/std Std 0.642703 +trainer/policy/normal/std Max 5.4562 +trainer/policy/normal/std Min 0.329271 +trainer/policy/normal/log_std Mean 0.992579 +trainer/policy/normal/log_std Std 0.302474 +trainer/policy/normal/log_std Max 1.69675 +trainer/policy/normal/log_std Min -1.11087 +eval/num steps total 780715 +eval/num paths total 783 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00336905 +eval/Actions Std 0.902151 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72978 +time/logging (s) 0.00396487 +time/sampling batch (s) 0.279383 +time/saving (s) 0.00400553 +time/training (s) 6.15435 +time/epoch (s) 9.17148 +time/total (s) 8107.62 +Epoch -218 +---------------------------------- --------------- +2022-05-10 15:26:06.138270 PDT | [1] Epoch -217 finished +---------------------------------- --------------- +epoch -217 +replay_buffer/size 999033 +trainer/num train calls 784000 +trainer/Policy Loss -20.6741 +trainer/Log Pis Mean 25.0663 +trainer/Log Pis Std 13.6071 +trainer/Log Pis Max 78.5934 +trainer/Log Pis Min -7.62091 +trainer/policy/mean Mean -0.0627406 +trainer/policy/mean Std 0.905779 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81518 +trainer/policy/normal/std Std 0.664111 +trainer/policy/normal/std Max 5.95989 +trainer/policy/normal/std Min 0.282644 +trainer/policy/normal/log_std Mean 0.994609 +trainer/policy/normal/log_std Std 0.321732 +trainer/policy/normal/log_std Max 1.78505 +trainer/policy/normal/log_std Min -1.26357 +eval/num steps total 781715 +eval/num paths total 784 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.392296 +eval/Actions Std 0.797055 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78153 +time/logging (s) 0.00366924 +time/sampling batch (s) 0.278387 +time/saving (s) 0.00343344 +time/training (s) 7.13476 +time/epoch (s) 10.2018 +time/total (s) 8117.83 +Epoch -217 +---------------------------------- --------------- +2022-05-10 15:26:16.030371 PDT | [1] Epoch -216 finished +---------------------------------- --------------- +epoch -216 +replay_buffer/size 999033 +trainer/num train calls 785000 +trainer/Policy Loss -19.0151 +trainer/Log Pis Mean 23.8631 +trainer/Log Pis Std 14.1812 +trainer/Log Pis Max 74.3014 +trainer/Log Pis Min -10.3293 +trainer/policy/mean Mean -0.0211728 +trainer/policy/mean Std 0.909673 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.84235 +trainer/policy/normal/std Std 0.669761 +trainer/policy/normal/std Max 6.91303 +trainer/policy/normal/std Min 0.26143 +trainer/policy/normal/log_std Mean 1.00527 +trainer/policy/normal/log_std Std 0.316302 +trainer/policy/normal/log_std Max 1.93341 +trainer/policy/normal/log_std Min -1.34159 +eval/num steps total 782715 +eval/num paths total 785 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0450492 +eval/Actions Std 0.908829 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57627 +time/logging (s) 0.00369538 +time/sampling batch (s) 0.276932 +time/saving (s) 0.00343984 +time/training (s) 7.00759 +time/epoch (s) 9.86792 +time/total (s) 8127.7 +Epoch -216 +---------------------------------- --------------- +2022-05-10 15:26:26.017967 PDT | [1] Epoch -215 finished +---------------------------------- --------------- +epoch -215 +replay_buffer/size 999033 +trainer/num train calls 786000 +trainer/Policy Loss -17.7053 +trainer/Log Pis Mean 23.3652 +trainer/Log Pis Std 13.4192 +trainer/Log Pis Max 67.3718 +trainer/Log Pis Min -6.91093 +trainer/policy/mean Mean -0.0466196 +trainer/policy/mean Std 0.900273 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81436 +trainer/policy/normal/std Std 0.669915 +trainer/policy/normal/std Max 6.00803 +trainer/policy/normal/std Min 0.286199 +trainer/policy/normal/log_std Mean 0.994416 +trainer/policy/normal/log_std Std 0.319614 +trainer/policy/normal/log_std Max 1.7931 +trainer/policy/normal/log_std Min -1.25107 +eval/num steps total 783715 +eval/num paths total 786 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0471942 +eval/Actions Std 0.906281 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57992 +time/logging (s) 0.00428026 +time/sampling batch (s) 0.279961 +time/saving (s) 0.00420312 +time/training (s) 7.09555 +time/epoch (s) 9.96391 +time/total (s) 8137.66 +Epoch -215 +---------------------------------- --------------- +2022-05-10 15:26:36.009530 PDT | [1] Epoch -214 finished +---------------------------------- --------------- +epoch -214 +replay_buffer/size 999033 +trainer/num train calls 787000 +trainer/Policy Loss -19.3952 +trainer/Log Pis Mean 24.2862 +trainer/Log Pis Std 12.496 +trainer/Log Pis Max 72.9971 +trainer/Log Pis Min -7.44744 +trainer/policy/mean Mean -0.0572861 +trainer/policy/mean Std 0.903343 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77766 +trainer/policy/normal/std Std 0.675312 +trainer/policy/normal/std Max 6.12022 +trainer/policy/normal/std Min 0.29497 +trainer/policy/normal/log_std Mean 0.978106 +trainer/policy/normal/log_std Std 0.334801 +trainer/policy/normal/log_std Max 1.8116 +trainer/policy/normal/log_std Min -1.22088 +eval/num steps total 784715 +eval/num paths total 787 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0850866 +eval/Actions Std 0.912495 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49472 +time/logging (s) 0.00415821 +time/sampling batch (s) 0.282997 +time/saving (s) 0.00408064 +time/training (s) 7.17982 +time/epoch (s) 9.96578 +time/total (s) 8147.63 +Epoch -214 +---------------------------------- --------------- +2022-05-10 15:26:46.634647 PDT | [1] Epoch -213 finished +---------------------------------- --------------- +epoch -213 +replay_buffer/size 999033 +trainer/num train calls 788000 +trainer/Policy Loss -19.9418 +trainer/Log Pis Mean 24.7067 +trainer/Log Pis Std 12.9694 +trainer/Log Pis Max 62.0927 +trainer/Log Pis Min -11.0661 +trainer/policy/mean Mean -0.0361634 +trainer/policy/mean Std 0.909564 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.831 +trainer/policy/normal/std Std 0.676362 +trainer/policy/normal/std Max 5.84817 +trainer/policy/normal/std Min 0.244379 +trainer/policy/normal/log_std Mean 0.997609 +trainer/policy/normal/log_std Std 0.336818 +trainer/policy/normal/log_std Max 1.76613 +trainer/policy/normal/log_std Min -1.40903 +eval/num steps total 785715 +eval/num paths total 788 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0298208 +eval/Actions Std 0.917444 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57954 +time/logging (s) 0.0038606 +time/sampling batch (s) 0.278282 +time/saving (s) 0.00353603 +time/training (s) 7.73468 +time/epoch (s) 10.5999 +time/total (s) 8158.24 +Epoch -213 +---------------------------------- --------------- +2022-05-10 15:26:57.449798 PDT | [1] Epoch -212 finished +---------------------------------- --------------- +epoch -212 +replay_buffer/size 999033 +trainer/num train calls 789000 +trainer/Policy Loss -19.7244 +trainer/Log Pis Mean 25.4856 +trainer/Log Pis Std 13.2955 +trainer/Log Pis Max 65.6471 +trainer/Log Pis Min -3.58707 +trainer/policy/mean Mean -0.0452864 +trainer/policy/mean Std 0.907284 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.86972 +trainer/policy/normal/std Std 0.694315 +trainer/policy/normal/std Max 6.51291 +trainer/policy/normal/std Min 0.23817 +trainer/policy/normal/log_std Mean 1.01212 +trainer/policy/normal/log_std Std 0.327803 +trainer/policy/normal/log_std Max 1.87379 +trainer/policy/normal/log_std Min -1.43477 +eval/num steps total 786715 +eval/num paths total 789 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0471026 +eval/Actions Std 0.915719 +eval/Actions Max 0.999997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55693 +time/logging (s) 0.00406618 +time/sampling batch (s) 0.52616 +time/saving (s) 0.00398173 +time/training (s) 7.69996 +time/epoch (s) 10.7911 +time/total (s) 8169.03 +Epoch -212 +---------------------------------- --------------- +2022-05-10 15:27:07.462720 PDT | [1] Epoch -211 finished +---------------------------------- --------------- +epoch -211 +replay_buffer/size 999033 +trainer/num train calls 790000 +trainer/Policy Loss -19.762 +trainer/Log Pis Mean 23.6071 +trainer/Log Pis Std 12.8254 +trainer/Log Pis Max 63.1834 +trainer/Log Pis Min -9.40792 +trainer/policy/mean Mean -0.0404392 +trainer/policy/mean Std 0.909076 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999975 +trainer/policy/normal/std Mean 2.76697 +trainer/policy/normal/std Std 0.659572 +trainer/policy/normal/std Max 5.98138 +trainer/policy/normal/std Min 0.27614 +trainer/policy/normal/log_std Mean 0.975558 +trainer/policy/normal/log_std Std 0.330506 +trainer/policy/normal/log_std Max 1.78865 +trainer/policy/normal/log_std Min -1.28685 +eval/num steps total 787715 +eval/num paths total 790 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.158599 +eval/Actions Std 0.899792 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55024 +time/logging (s) 0.00368981 +time/sampling batch (s) 0.275807 +time/saving (s) 0.00342827 +time/training (s) 7.15501 +time/epoch (s) 9.98818 +time/total (s) 8179.02 +Epoch -211 +---------------------------------- --------------- +2022-05-10 15:27:18.049612 PDT | [1] Epoch -210 finished +---------------------------------- --------------- +epoch -210 +replay_buffer/size 999033 +trainer/num train calls 791000 +trainer/Policy Loss -19.6191 +trainer/Log Pis Mean 23.8388 +trainer/Log Pis Std 13.3121 +trainer/Log Pis Max 69.0139 +trainer/Log Pis Min -9.21495 +trainer/policy/mean Mean -0.0246634 +trainer/policy/mean Std 0.906758 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82838 +trainer/policy/normal/std Std 0.689059 +trainer/policy/normal/std Max 6.07659 +trainer/policy/normal/std Min 0.294369 +trainer/policy/normal/log_std Mean 0.995623 +trainer/policy/normal/log_std Std 0.3384 +trainer/policy/normal/log_std Max 1.80444 +trainer/policy/normal/log_std Min -1.22292 +eval/num steps total 788715 +eval/num paths total 791 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.157683 +eval/Actions Std 0.8932 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6159 +time/logging (s) 0.00419955 +time/sampling batch (s) 0.532415 +time/saving (s) 0.00382135 +time/training (s) 7.40689 +time/epoch (s) 10.5632 +time/total (s) 8189.59 +Epoch -210 +---------------------------------- --------------- +2022-05-10 15:27:28.720337 PDT | [1] Epoch -209 finished +---------------------------------- --------------- +epoch -209 +replay_buffer/size 999033 +trainer/num train calls 792000 +trainer/Policy Loss -20.8624 +trainer/Log Pis Mean 24.4054 +trainer/Log Pis Std 13.6707 +trainer/Log Pis Max 76.3729 +trainer/Log Pis Min -6.68448 +trainer/policy/mean Mean -0.0433875 +trainer/policy/mean Std 0.905123 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81166 +trainer/policy/normal/std Std 0.681247 +trainer/policy/normal/std Max 5.77115 +trainer/policy/normal/std Min 0.275566 +trainer/policy/normal/log_std Mean 0.991669 +trainer/policy/normal/log_std Std 0.327217 +trainer/policy/normal/log_std Max 1.75287 +trainer/policy/normal/log_std Min -1.28893 +eval/num steps total 789715 +eval/num paths total 792 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0113505 +eval/Actions Std 0.903679 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54859 +time/logging (s) 0.0039364 +time/sampling batch (s) 0.282419 +time/saving (s) 0.00369046 +time/training (s) 7.80699 +time/epoch (s) 10.6456 +time/total (s) 8200.24 +Epoch -209 +---------------------------------- --------------- +2022-05-10 15:27:39.091683 PDT | [1] Epoch -208 finished +---------------------------------- --------------- +epoch -208 +replay_buffer/size 999033 +trainer/num train calls 793000 +trainer/Policy Loss -20.3976 +trainer/Log Pis Mean 24.1403 +trainer/Log Pis Std 13.534 +trainer/Log Pis Max 66.4486 +trainer/Log Pis Min -10.7337 +trainer/policy/mean Mean -0.0450693 +trainer/policy/mean Std 0.910401 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80933 +trainer/policy/normal/std Std 0.645904 +trainer/policy/normal/std Max 6.51164 +trainer/policy/normal/std Min 0.22122 +trainer/policy/normal/log_std Mean 0.993881 +trainer/policy/normal/log_std Std 0.31922 +trainer/policy/normal/log_std Max 1.87359 +trainer/policy/normal/log_std Min -1.5086 +eval/num steps total 790715 +eval/num paths total 793 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.221752 +eval/Actions Std 0.86552 +eval/Actions Max 0.999998 +eval/Actions Min -0.999942 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63067 +time/logging (s) 0.0037868 +time/sampling batch (s) 0.528651 +time/saving (s) 0.00345293 +time/training (s) 7.1803 +time/epoch (s) 10.3469 +time/total (s) 8210.59 +Epoch -208 +---------------------------------- --------------- +2022-05-10 15:27:48.857867 PDT | [1] Epoch -207 finished +---------------------------------- --------------- +epoch -207 +replay_buffer/size 999033 +trainer/num train calls 794000 +trainer/Policy Loss -19.4086 +trainer/Log Pis Mean 24.8826 +trainer/Log Pis Std 12.8719 +trainer/Log Pis Max 63.3763 +trainer/Log Pis Min -6.40666 +trainer/policy/mean Mean -0.045766 +trainer/policy/mean Std 0.904868 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89022 +trainer/policy/normal/std Std 0.663607 +trainer/policy/normal/std Max 5.77961 +trainer/policy/normal/std Min 0.309846 +trainer/policy/normal/log_std Mean 1.02433 +trainer/policy/normal/log_std Std 0.304844 +trainer/policy/normal/log_std Max 1.75434 +trainer/policy/normal/log_std Min -1.17168 +eval/num steps total 791715 +eval/num paths total 794 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.155735 +eval/Actions Std 0.866059 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56196 +time/logging (s) 0.00370096 +time/sampling batch (s) 0.277455 +time/saving (s) 0.0034406 +time/training (s) 6.89527 +time/epoch (s) 9.74182 +time/total (s) 8220.33 +Epoch -207 +---------------------------------- --------------- +2022-05-10 15:27:59.216462 PDT | [1] Epoch -206 finished +---------------------------------- --------------- +epoch -206 +replay_buffer/size 999033 +trainer/num train calls 795000 +trainer/Policy Loss -20.1039 +trainer/Log Pis Mean 23.3536 +trainer/Log Pis Std 13.061 +trainer/Log Pis Max 79.114 +trainer/Log Pis Min -9.76407 +trainer/policy/mean Mean -0.0184689 +trainer/policy/mean Std 0.906143 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7694 +trainer/policy/normal/std Std 0.64729 +trainer/policy/normal/std Max 5.95984 +trainer/policy/normal/std Min 0.246149 +trainer/policy/normal/log_std Mean 0.980286 +trainer/policy/normal/log_std Std 0.310908 +trainer/policy/normal/log_std Max 1.78504 +trainer/policy/normal/log_std Min -1.40182 +eval/num steps total 792715 +eval/num paths total 795 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141822 +eval/Actions Std 0.917924 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65522 +time/logging (s) 0.00369143 +time/sampling batch (s) 0.278358 +time/saving (s) 0.00345911 +time/training (s) 7.39341 +time/epoch (s) 10.3341 +time/total (s) 8230.67 +Epoch -206 +---------------------------------- --------------- +2022-05-10 15:28:10.797674 PDT | [1] Epoch -205 finished +---------------------------------- --------------- +epoch -205 +replay_buffer/size 999033 +trainer/num train calls 796000 +trainer/Policy Loss -18.3139 +trainer/Log Pis Mean 24.4705 +trainer/Log Pis Std 12.7652 +trainer/Log Pis Max 64.9137 +trainer/Log Pis Min -6.02879 +trainer/policy/mean Mean -0.0352178 +trainer/policy/mean Std 0.902666 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80048 +trainer/policy/normal/std Std 0.659872 +trainer/policy/normal/std Max 5.78477 +trainer/policy/normal/std Min 0.265455 +trainer/policy/normal/log_std Mean 0.988951 +trainer/policy/normal/log_std Std 0.324727 +trainer/policy/normal/log_std Max 1.75523 +trainer/policy/normal/log_std Min -1.32631 +eval/num steps total 793715 +eval/num paths total 796 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.048975 +eval/Actions Std 0.885921 +eval/Actions Max 0.999982 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60562 +time/logging (s) 0.00409774 +time/sampling batch (s) 0.531568 +time/saving (s) 0.00383172 +time/training (s) 8.41204 +time/epoch (s) 11.5572 +time/total (s) 8242.23 +Epoch -205 +---------------------------------- --------------- +2022-05-10 15:28:22.014740 PDT | [1] Epoch -204 finished +---------------------------------- --------------- +epoch -204 +replay_buffer/size 999033 +trainer/num train calls 797000 +trainer/Policy Loss -19.2523 +trainer/Log Pis Mean 24.047 +trainer/Log Pis Std 13.5248 +trainer/Log Pis Max 62.2163 +trainer/Log Pis Min -6.64864 +trainer/policy/mean Mean -0.0405101 +trainer/policy/mean Std 0.906203 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79003 +trainer/policy/normal/std Std 0.666223 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.185217 +trainer/policy/normal/log_std Mean 0.984571 +trainer/policy/normal/log_std Std 0.326919 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.68623 +eval/num steps total 794715 +eval/num paths total 797 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.170748 +eval/Actions Std 0.91595 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89161 +time/logging (s) 0.00394556 +time/sampling batch (s) 0.579087 +time/saving (s) 0.00396065 +time/training (s) 7.71298 +time/epoch (s) 11.1916 +time/total (s) 8253.42 +Epoch -204 +---------------------------------- --------------- +2022-05-10 15:28:33.766331 PDT | [1] Epoch -203 finished +---------------------------------- --------------- +epoch -203 +replay_buffer/size 999033 +trainer/num train calls 798000 +trainer/Policy Loss -20.2208 +trainer/Log Pis Mean 25.2008 +trainer/Log Pis Std 13.0435 +trainer/Log Pis Max 64.7555 +trainer/Log Pis Min -11.6704 +trainer/policy/mean Mean -0.0280698 +trainer/policy/mean Std 0.909728 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.75401 +trainer/policy/normal/std Std 0.658712 +trainer/policy/normal/std Max 7.12054 +trainer/policy/normal/std Min 0.244007 +trainer/policy/normal/log_std Mean 0.971039 +trainer/policy/normal/log_std Std 0.330114 +trainer/policy/normal/log_std Max 1.96298 +trainer/policy/normal/log_std Min -1.41056 +eval/num steps total 795715 +eval/num paths total 798 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0150148 +eval/Actions Std 0.90854 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6595 +time/logging (s) 0.00388691 +time/sampling batch (s) 1.08264 +time/saving (s) 0.00364699 +time/training (s) 7.9767 +time/epoch (s) 11.7264 +time/total (s) 8265.15 +Epoch -203 +---------------------------------- --------------- +2022-05-10 15:28:45.243023 PDT | [1] Epoch -202 finished +---------------------------------- --------------- +epoch -202 +replay_buffer/size 999033 +trainer/num train calls 799000 +trainer/Policy Loss -18.5306 +trainer/Log Pis Mean 24.5223 +trainer/Log Pis Std 13.6251 +trainer/Log Pis Max 74.6016 +trainer/Log Pis Min -7.35689 +trainer/policy/mean Mean -0.0270938 +trainer/policy/mean Std 0.90388 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77171 +trainer/policy/normal/std Std 0.677253 +trainer/policy/normal/std Max 6.09195 +trainer/policy/normal/std Min 0.236926 +trainer/policy/normal/log_std Mean 0.973827 +trainer/policy/normal/log_std Std 0.347621 +trainer/policy/normal/log_std Max 1.80697 +trainer/policy/normal/log_std Min -1.44001 +eval/num steps total 796715 +eval/num paths total 799 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0964336 +eval/Actions Std 0.918281 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88178 +time/logging (s) 0.00380402 +time/sampling batch (s) 0.58051 +time/saving (s) 0.0036647 +time/training (s) 7.98179 +time/epoch (s) 11.4515 +time/total (s) 8276.61 +Epoch -202 +---------------------------------- --------------- +2022-05-10 15:28:56.920783 PDT | [1] Epoch -201 finished +---------------------------------- --------------- +epoch -201 +replay_buffer/size 999033 +trainer/num train calls 800000 +trainer/Policy Loss -19.4717 +trainer/Log Pis Mean 23.5678 +trainer/Log Pis Std 12.8617 +trainer/Log Pis Max 63.6778 +trainer/Log Pis Min -7.54808 +trainer/policy/mean Mean -0.0175277 +trainer/policy/mean Std 0.9042 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.75854 +trainer/policy/normal/std Std 0.665237 +trainer/policy/normal/std Max 5.04577 +trainer/policy/normal/std Min 0.256719 +trainer/policy/normal/log_std Mean 0.970462 +trainer/policy/normal/log_std Std 0.340742 +trainer/policy/normal/log_std Max 1.61855 +trainer/policy/normal/log_std Min -1.35977 +eval/num steps total 797715 +eval/num paths total 800 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.367508 +eval/Actions Std 0.853137 +eval/Actions Max 0.999968 +eval/Actions Min -0.999949 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.9866 +time/logging (s) 0.00377999 +time/sampling batch (s) 0.336675 +time/saving (s) 0.00677541 +time/training (s) 8.31849 +time/epoch (s) 11.6523 +time/total (s) 8288.26 +Epoch -201 +---------------------------------- --------------- +2022-05-10 15:29:08.340321 PDT | [1] Epoch -200 finished +---------------------------------- --------------- +epoch -200 +replay_buffer/size 999033 +trainer/num train calls 801000 +trainer/Policy Loss -19.1487 +trainer/Log Pis Mean 23.7851 +trainer/Log Pis Std 13.2635 +trainer/Log Pis Max 61.3205 +trainer/Log Pis Min -7.7625 +trainer/policy/mean Mean -0.0419203 +trainer/policy/mean Std 0.9049 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7862 +trainer/policy/normal/std Std 0.683741 +trainer/policy/normal/std Max 5.82508 +trainer/policy/normal/std Min 0.261349 +trainer/policy/normal/log_std Mean 0.980703 +trainer/policy/normal/log_std Std 0.337013 +trainer/policy/normal/log_std Max 1.76217 +trainer/policy/normal/log_std Min -1.3419 +eval/num steps total 798715 +eval/num paths total 801 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0382091 +eval/Actions Std 0.913274 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80569 +time/logging (s) 0.00426042 +time/sampling batch (s) 0.830597 +time/saving (s) 0.00401079 +time/training (s) 7.74983 +time/epoch (s) 11.3944 +time/total (s) 8299.66 +Epoch -200 +---------------------------------- --------------- +2022-05-10 15:29:18.907178 PDT | [1] Epoch -199 finished +---------------------------------- --------------- +epoch -199 +replay_buffer/size 999033 +trainer/num train calls 802000 +trainer/Policy Loss -19.5095 +trainer/Log Pis Mean 23.8217 +trainer/Log Pis Std 12.5777 +trainer/Log Pis Max 64.3232 +trainer/Log Pis Min -6.76637 +trainer/policy/mean Mean -0.0390414 +trainer/policy/mean Std 0.909027 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80131 +trainer/policy/normal/std Std 0.667254 +trainer/policy/normal/std Max 6.18311 +trainer/policy/normal/std Min 0.239116 +trainer/policy/normal/log_std Mean 0.988187 +trainer/policy/normal/log_std Std 0.329459 +trainer/policy/normal/log_std Max 1.82182 +trainer/policy/normal/log_std Min -1.4308 +eval/num steps total 799715 +eval/num paths total 802 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0826737 +eval/Actions Std 0.879068 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60625 +time/logging (s) 0.0037722 +time/sampling batch (s) 0.280899 +time/saving (s) 0.00359337 +time/training (s) 7.64677 +time/epoch (s) 10.5413 +time/total (s) 8310.21 +Epoch -199 +---------------------------------- --------------- +2022-05-10 15:29:30.461395 PDT | [1] Epoch -198 finished +---------------------------------- --------------- +epoch -198 +replay_buffer/size 999033 +trainer/num train calls 803000 +trainer/Policy Loss -19.341 +trainer/Log Pis Mean 26.1857 +trainer/Log Pis Std 13.0749 +trainer/Log Pis Max 70.1787 +trainer/Log Pis Min -4.96213 +trainer/policy/mean Mean -0.0237945 +trainer/policy/mean Std 0.913582 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80326 +trainer/policy/normal/std Std 0.632985 +trainer/policy/normal/std Max 5.59829 +trainer/policy/normal/std Min 0.248264 +trainer/policy/normal/log_std Mean 0.993978 +trainer/policy/normal/log_std Std 0.307132 +trainer/policy/normal/log_std Max 1.72246 +trainer/policy/normal/log_std Min -1.39326 +eval/num steps total 800715 +eval/num paths total 803 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.449059 +eval/Actions Std 0.704849 +eval/Actions Max 0.999995 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58491 +time/logging (s) 0.00365573 +time/sampling batch (s) 0.276933 +time/saving (s) 0.00333751 +time/training (s) 8.66097 +time/epoch (s) 11.5298 +time/total (s) 8321.74 +Epoch -198 +---------------------------------- --------------- +2022-05-10 15:29:40.642295 PDT | [1] Epoch -197 finished +---------------------------------- --------------- +epoch -197 +replay_buffer/size 999033 +trainer/num train calls 804000 +trainer/Policy Loss -20.269 +trainer/Log Pis Mean 24.983 +trainer/Log Pis Std 13.1599 +trainer/Log Pis Max 63.9249 +trainer/Log Pis Min -6.82395 +trainer/policy/mean Mean -0.0557731 +trainer/policy/mean Std 0.907163 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8824 +trainer/policy/normal/std Std 0.662932 +trainer/policy/normal/std Max 5.60918 +trainer/policy/normal/std Min 0.257326 +trainer/policy/normal/log_std Mean 1.01941 +trainer/policy/normal/log_std Std 0.318923 +trainer/policy/normal/log_std Max 1.7244 +trainer/policy/normal/log_std Min -1.35741 +eval/num steps total 801715 +eval/num paths total 804 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00131638 +eval/Actions Std 0.912331 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57063 +time/logging (s) 0.00379842 +time/sampling batch (s) 0.525418 +time/saving (s) 0.00349235 +time/training (s) 7.05376 +time/epoch (s) 10.1571 +time/total (s) 8331.9 +Epoch -197 +---------------------------------- --------------- +2022-05-10 15:29:50.119682 PDT | [1] Epoch -196 finished +---------------------------------- --------------- +epoch -196 +replay_buffer/size 999033 +trainer/num train calls 805000 +trainer/Policy Loss -19.706 +trainer/Log Pis Mean 24.1062 +trainer/Log Pis Std 13.6331 +trainer/Log Pis Max 78.394 +trainer/Log Pis Min -3.7386 +trainer/policy/mean Mean -0.0502013 +trainer/policy/mean Std 0.903294 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81757 +trainer/policy/normal/std Std 0.698961 +trainer/policy/normal/std Max 5.69979 +trainer/policy/normal/std Min 0.234908 +trainer/policy/normal/log_std Mean 0.990262 +trainer/policy/normal/log_std Std 0.34361 +trainer/policy/normal/log_std Max 1.74043 +trainer/policy/normal/log_std Min -1.44856 +eval/num steps total 802715 +eval/num paths total 805 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.170446 +eval/Actions Std 0.920028 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55997 +time/logging (s) 0.00427797 +time/sampling batch (s) 0.283358 +time/saving (s) 0.00467063 +time/training (s) 6.60057 +time/epoch (s) 9.45284 +time/total (s) 8341.35 +Epoch -196 +---------------------------------- --------------- +2022-05-10 15:30:00.659897 PDT | [1] Epoch -195 finished +---------------------------------- --------------- +epoch -195 +replay_buffer/size 999033 +trainer/num train calls 806000 +trainer/Policy Loss -19.338 +trainer/Log Pis Mean 24.1391 +trainer/Log Pis Std 13.2928 +trainer/Log Pis Max 63.1859 +trainer/Log Pis Min -6.46183 +trainer/policy/mean Mean -0.0366744 +trainer/policy/mean Std 0.900149 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8341 +trainer/policy/normal/std Std 0.693314 +trainer/policy/normal/std Max 5.8304 +trainer/policy/normal/std Min 0.234398 +trainer/policy/normal/log_std Mean 0.997133 +trainer/policy/normal/log_std Std 0.34193 +trainer/policy/normal/log_std Max 1.76309 +trainer/policy/normal/log_std Min -1.45073 +eval/num steps total 803715 +eval/num paths total 806 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0753911 +eval/Actions Std 0.909825 +eval/Actions Max 0.999966 +eval/Actions Min -0.999974 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56627 +time/logging (s) 0.00429186 +time/sampling batch (s) 0.282976 +time/saving (s) 0.004125 +time/training (s) 7.65596 +time/epoch (s) 10.5136 +time/total (s) 8351.87 +Epoch -195 +---------------------------------- --------------- +2022-05-10 15:30:11.416069 PDT | [1] Epoch -194 finished +---------------------------------- --------------- +epoch -194 +replay_buffer/size 999033 +trainer/num train calls 807000 +trainer/Policy Loss -19.7073 +trainer/Log Pis Mean 24.3984 +trainer/Log Pis Std 12.9635 +trainer/Log Pis Max 68.0666 +trainer/Log Pis Min -11.7807 +trainer/policy/mean Mean -0.0424841 +trainer/policy/mean Std 0.905254 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79277 +trainer/policy/normal/std Std 0.657195 +trainer/policy/normal/std Max 6.007 +trainer/policy/normal/std Min 0.303884 +trainer/policy/normal/log_std Mean 0.986886 +trainer/policy/normal/log_std Std 0.321188 +trainer/policy/normal/log_std Max 1.79293 +trainer/policy/normal/log_std Min -1.19111 +eval/num steps total 804715 +eval/num paths total 807 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118363 +eval/Actions Std 0.892089 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50409 +time/logging (s) 0.00426347 +time/sampling batch (s) 0.534292 +time/saving (s) 0.00392724 +time/training (s) 7.6836 +time/epoch (s) 10.7302 +time/total (s) 8362.61 +Epoch -194 +---------------------------------- --------------- +2022-05-10 15:30:22.212750 PDT | [1] Epoch -193 finished +---------------------------------- --------------- +epoch -193 +replay_buffer/size 999033 +trainer/num train calls 808000 +trainer/Policy Loss -19.3669 +trainer/Log Pis Mean 25.2471 +trainer/Log Pis Std 13.6818 +trainer/Log Pis Max 90.4427 +trainer/Log Pis Min -8.90776 +trainer/policy/mean Mean -0.0472313 +trainer/policy/mean Std 0.906329 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.76898 +trainer/policy/normal/std Std 0.672809 +trainer/policy/normal/std Max 6.07261 +trainer/policy/normal/std Min 0.315336 +trainer/policy/normal/log_std Mean 0.974852 +trainer/policy/normal/log_std Std 0.335592 +trainer/policy/normal/log_std Max 1.80379 +trainer/policy/normal/log_std Min -1.15412 +eval/num steps total 805715 +eval/num paths total 808 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0870496 +eval/Actions Std 0.913253 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65144 +time/logging (s) 0.00379375 +time/sampling batch (s) 0.526977 +time/saving (s) 0.00353927 +time/training (s) 7.58569 +time/epoch (s) 10.7714 +time/total (s) 8373.38 +Epoch -193 +---------------------------------- --------------- +2022-05-10 15:30:32.089196 PDT | [1] Epoch -192 finished +---------------------------------- --------------- +epoch -192 +replay_buffer/size 999033 +trainer/num train calls 809000 +trainer/Policy Loss -20.0539 +trainer/Log Pis Mean 23.95 +trainer/Log Pis Std 12.7266 +trainer/Log Pis Max 68.8137 +trainer/Log Pis Min -4.99732 +trainer/policy/mean Mean -0.0438201 +trainer/policy/mean Std 0.904701 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81667 +trainer/policy/normal/std Std 0.682084 +trainer/policy/normal/std Max 5.53725 +trainer/policy/normal/std Min 0.314194 +trainer/policy/normal/log_std Mean 0.991952 +trainer/policy/normal/log_std Std 0.336481 +trainer/policy/normal/log_std Max 1.7115 +trainer/policy/normal/log_std Min -1.15774 +eval/num steps total 806715 +eval/num paths total 809 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.182997 +eval/Actions Std 0.9455 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6187 +time/logging (s) 0.00371471 +time/sampling batch (s) 0.279888 +time/saving (s) 0.00341384 +time/training (s) 6.94627 +time/epoch (s) 9.85199 +time/total (s) 8383.24 +Epoch -192 +---------------------------------- --------------- +2022-05-10 15:30:42.184433 PDT | [1] Epoch -191 finished +---------------------------------- --------------- +epoch -191 +replay_buffer/size 999033 +trainer/num train calls 810000 +trainer/Policy Loss -19.4648 +trainer/Log Pis Mean 22.64 +trainer/Log Pis Std 13.0509 +trainer/Log Pis Max 73.892 +trainer/Log Pis Min -10.3257 +trainer/policy/mean Mean 0.00370978 +trainer/policy/mean Std 0.904437 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.72469 +trainer/policy/normal/std Std 0.667063 +trainer/policy/normal/std Max 6.30757 +trainer/policy/normal/std Min 0.241705 +trainer/policy/normal/log_std Mean 0.958094 +trainer/policy/normal/log_std Std 0.338672 +trainer/policy/normal/log_std Max 1.84175 +trainer/policy/normal/log_std Min -1.42004 +eval/num steps total 807715 +eval/num paths total 810 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.442852 +eval/Actions Std 0.775026 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63614 +time/logging (s) 0.00421165 +time/sampling batch (s) 0.278109 +time/saving (s) 0.00397377 +time/training (s) 7.14883 +time/epoch (s) 10.0713 +time/total (s) 8393.31 +Epoch -191 +---------------------------------- --------------- +2022-05-10 15:30:53.055041 PDT | [1] Epoch -190 finished +---------------------------------- --------------- +epoch -190 +replay_buffer/size 999033 +trainer/num train calls 811000 +trainer/Policy Loss -19.5712 +trainer/Log Pis Mean 25.0529 +trainer/Log Pis Std 13.6829 +trainer/Log Pis Max 63.0665 +trainer/Log Pis Min -8.23893 +trainer/policy/mean Mean -0.056302 +trainer/policy/mean Std 0.90765 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83011 +trainer/policy/normal/std Std 0.679234 +trainer/policy/normal/std Max 6.15288 +trainer/policy/normal/std Min 0.221917 +trainer/policy/normal/log_std Mean 0.997322 +trainer/policy/normal/log_std Std 0.334484 +trainer/policy/normal/log_std Max 1.81692 +trainer/policy/normal/log_std Min -1.50545 +eval/num steps total 808715 +eval/num paths total 811 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.134926 +eval/Actions Std 0.86786 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70327 +time/logging (s) 0.00374036 +time/sampling batch (s) 0.278652 +time/saving (s) 0.00354755 +time/training (s) 7.8558 +time/epoch (s) 10.845 +time/total (s) 8404.16 +Epoch -190 +---------------------------------- --------------- +2022-05-10 15:31:04.256093 PDT | [1] Epoch -189 finished +---------------------------------- --------------- +epoch -189 +replay_buffer/size 999033 +trainer/num train calls 812000 +trainer/Policy Loss -19.3262 +trainer/Log Pis Mean 24.9678 +trainer/Log Pis Std 13.1654 +trainer/Log Pis Max 70.4626 +trainer/Log Pis Min -8.94855 +trainer/policy/mean Mean -0.0470947 +trainer/policy/mean Std 0.906593 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77179 +trainer/policy/normal/std Std 0.676779 +trainer/policy/normal/std Max 6.76118 +trainer/policy/normal/std Min 0.283447 +trainer/policy/normal/log_std Mean 0.975391 +trainer/policy/normal/log_std Std 0.337983 +trainer/policy/normal/log_std Max 1.9112 +trainer/policy/normal/log_std Min -1.26073 +eval/num steps total 809715 +eval/num paths total 812 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.113681 +eval/Actions Std 0.875196 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5826 +time/logging (s) 0.00426317 +time/sampling batch (s) 0.279556 +time/saving (s) 0.00344754 +time/training (s) 8.30703 +time/epoch (s) 11.1769 +time/total (s) 8415.34 +Epoch -189 +---------------------------------- --------------- +2022-05-10 15:31:14.929326 PDT | [1] Epoch -188 finished +---------------------------------- --------------- +epoch -188 +replay_buffer/size 999033 +trainer/num train calls 813000 +trainer/Policy Loss -18.8717 +trainer/Log Pis Mean 24.4178 +trainer/Log Pis Std 12.17 +trainer/Log Pis Max 57.0086 +trainer/Log Pis Min -5.24793 +trainer/policy/mean Mean -0.0243428 +trainer/policy/mean Std 0.904148 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81065 +trainer/policy/normal/std Std 0.666783 +trainer/policy/normal/std Max 6.3337 +trainer/policy/normal/std Min 0.339031 +trainer/policy/normal/log_std Mean 0.991896 +trainer/policy/normal/log_std Std 0.327172 +trainer/policy/normal/log_std Max 1.84588 +trainer/policy/normal/log_std Min -1.08166 +eval/num steps total 810715 +eval/num paths total 813 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00412394 +eval/Actions Std 0.898311 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6754 +time/logging (s) 0.00368659 +time/sampling batch (s) 0.276425 +time/saving (s) 0.00340251 +time/training (s) 7.68906 +time/epoch (s) 10.648 +time/total (s) 8425.99 +Epoch -188 +---------------------------------- --------------- +2022-05-10 15:31:24.642946 PDT | [1] Epoch -187 finished +---------------------------------- --------------- +epoch -187 +replay_buffer/size 999033 +trainer/num train calls 814000 +trainer/Policy Loss -18.8655 +trainer/Log Pis Mean 24.4876 +trainer/Log Pis Std 13.3315 +trainer/Log Pis Max 69.1731 +trainer/Log Pis Min -7.95466 +trainer/policy/mean Mean -0.0557277 +trainer/policy/mean Std 0.904795 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83253 +trainer/policy/normal/std Std 0.659095 +trainer/policy/normal/std Max 5.82331 +trainer/policy/normal/std Min 0.261869 +trainer/policy/normal/log_std Mean 1.00271 +trainer/policy/normal/log_std Std 0.312805 +trainer/policy/normal/log_std Max 1.76187 +trainer/policy/normal/log_std Min -1.33991 +eval/num steps total 811715 +eval/num paths total 814 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0720447 +eval/Actions Std 0.883529 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71036 +time/logging (s) 0.0037528 +time/sampling batch (s) 0.527769 +time/saving (s) 0.00340959 +time/training (s) 6.44401 +time/epoch (s) 9.68931 +time/total (s) 8435.68 +Epoch -187 +---------------------------------- --------------- +2022-05-10 15:31:36.260022 PDT | [1] Epoch -186 finished +---------------------------------- --------------- +epoch -186 +replay_buffer/size 999033 +trainer/num train calls 815000 +trainer/Policy Loss -19.0671 +trainer/Log Pis Mean 23.8852 +trainer/Log Pis Std 13.1484 +trainer/Log Pis Max 67.8324 +trainer/Log Pis Min -11.8889 +trainer/policy/mean Mean -0.0343303 +trainer/policy/mean Std 0.902989 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80809 +trainer/policy/normal/std Std 0.683997 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.264301 +trainer/policy/normal/log_std Mean 0.989368 +trainer/policy/normal/log_std Std 0.333722 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.33067 +eval/num steps total 812715 +eval/num paths total 815 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118802 +eval/Actions Std 0.911729 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58253 +time/logging (s) 0.00405638 +time/sampling batch (s) 0.527836 +time/saving (s) 0.00390798 +time/training (s) 8.47454 +time/epoch (s) 11.5929 +time/total (s) 8447.28 +Epoch -186 +---------------------------------- --------------- +2022-05-10 15:31:46.976758 PDT | [1] Epoch -185 finished +---------------------------------- --------------- +epoch -185 +replay_buffer/size 999033 +trainer/num train calls 816000 +trainer/Policy Loss -19.9017 +trainer/Log Pis Mean 24.2739 +trainer/Log Pis Std 13.403 +trainer/Log Pis Max 68.4424 +trainer/Log Pis Min -10.6758 +trainer/policy/mean Mean -0.0334365 +trainer/policy/mean Std 0.908077 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79433 +trainer/policy/normal/std Std 0.686002 +trainer/policy/normal/std Max 5.34573 +trainer/policy/normal/std Min 0.255835 +trainer/policy/normal/log_std Mean 0.980887 +trainer/policy/normal/log_std Std 0.353157 +trainer/policy/normal/log_std Max 1.6763 +trainer/policy/normal/log_std Min -1.36322 +eval/num steps total 813715 +eval/num paths total 816 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.418415 +eval/Actions Std 0.829064 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44312 +time/logging (s) 0.00377687 +time/sampling batch (s) 0.279744 +time/saving (s) 0.00355522 +time/training (s) 7.71086 +time/epoch (s) 10.4411 +time/total (s) 8457.97 +Epoch -185 +---------------------------------- --------------- +2022-05-10 15:31:58.422525 PDT | [1] Epoch -184 finished +---------------------------------- --------------- +epoch -184 +replay_buffer/size 999033 +trainer/num train calls 817000 +trainer/Policy Loss -19.4257 +trainer/Log Pis Mean 24.7389 +trainer/Log Pis Std 12.9719 +trainer/Log Pis Max 74.6039 +trainer/Log Pis Min -7.17021 +trainer/policy/mean Mean -0.0304705 +trainer/policy/mean Std 0.906293 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84812 +trainer/policy/normal/std Std 0.660733 +trainer/policy/normal/std Max 5.66188 +trainer/policy/normal/std Min 0.282226 +trainer/policy/normal/log_std Mean 1.00704 +trainer/policy/normal/log_std Std 0.320757 +trainer/policy/normal/log_std Max 1.73376 +trainer/policy/normal/log_std Min -1.26505 +eval/num steps total 814715 +eval/num paths total 817 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101004 +eval/Actions Std 0.919536 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75887 +time/logging (s) 0.0039251 +time/sampling batch (s) 0.285109 +time/saving (s) 0.00348805 +time/training (s) 8.36955 +time/epoch (s) 11.4209 +time/total (s) 8469.4 +Epoch -184 +---------------------------------- --------------- +2022-05-10 15:32:08.606643 PDT | [1] Epoch -183 finished +---------------------------------- --------------- +epoch -183 +replay_buffer/size 999033 +trainer/num train calls 818000 +trainer/Policy Loss -20.7531 +trainer/Log Pis Mean 24.5316 +trainer/Log Pis Std 13.133 +trainer/Log Pis Max 68.9731 +trainer/Log Pis Min -5.44341 +trainer/policy/mean Mean -0.0548179 +trainer/policy/mean Std 0.903828 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78288 +trainer/policy/normal/std Std 0.668548 +trainer/policy/normal/std Max 5.62457 +trainer/policy/normal/std Min 0.250564 +trainer/policy/normal/log_std Mean 0.98085 +trainer/policy/normal/log_std Std 0.332556 +trainer/policy/normal/log_std Max 1.72714 +trainer/policy/normal/log_std Min -1.38404 +eval/num steps total 815715 +eval/num paths total 818 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0188453 +eval/Actions Std 0.900034 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78804 +time/logging (s) 0.00378876 +time/sampling batch (s) 0.533136 +time/saving (s) 0.00345304 +time/training (s) 6.83093 +time/epoch (s) 10.1594 +time/total (s) 8479.56 +Epoch -183 +---------------------------------- --------------- +2022-05-10 15:32:18.929301 PDT | [1] Epoch -182 finished +---------------------------------- --------------- +epoch -182 +replay_buffer/size 999033 +trainer/num train calls 819000 +trainer/Policy Loss -19.1109 +trainer/Log Pis Mean 24.0307 +trainer/Log Pis Std 13.304 +trainer/Log Pis Max 65.0672 +trainer/Log Pis Min -8.52987 +trainer/policy/mean Mean -0.0356447 +trainer/policy/mean Std 0.900769 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76583 +trainer/policy/normal/std Std 0.691096 +trainer/policy/normal/std Max 5.58527 +trainer/policy/normal/std Min 0.20906 +trainer/policy/normal/log_std Mean 0.972252 +trainer/policy/normal/log_std Std 0.340255 +trainer/policy/normal/log_std Max 1.72013 +trainer/policy/normal/log_std Min -1.56513 +eval/num steps total 816715 +eval/num paths total 819 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119522 +eval/Actions Std 0.88672 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72751 +time/logging (s) 0.00372667 +time/sampling batch (s) 0.780203 +time/saving (s) 0.00344554 +time/training (s) 6.78331 +time/epoch (s) 10.2982 +time/total (s) 8489.86 +Epoch -182 +---------------------------------- --------------- +2022-05-10 15:32:30.422939 PDT | [1] Epoch -181 finished +---------------------------------- --------------- +epoch -181 +replay_buffer/size 999033 +trainer/num train calls 820000 +trainer/Policy Loss -19.754 +trainer/Log Pis Mean 24.3655 +trainer/Log Pis Std 13.4284 +trainer/Log Pis Max 77.8238 +trainer/Log Pis Min -8.3664 +trainer/policy/mean Mean -0.0296693 +trainer/policy/mean Std 0.911296 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78876 +trainer/policy/normal/std Std 0.663446 +trainer/policy/normal/std Max 5.38617 +trainer/policy/normal/std Min 0.281595 +trainer/policy/normal/log_std Mean 0.983616 +trainer/policy/normal/log_std Std 0.329732 +trainer/policy/normal/log_std Max 1.68383 +trainer/policy/normal/log_std Min -1.26728 +eval/num steps total 817715 +eval/num paths total 820 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112204 +eval/Actions Std 0.892254 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62333 +time/logging (s) 0.00385251 +time/sampling batch (s) 0.529709 +time/saving (s) 0.00340094 +time/training (s) 8.30885 +time/epoch (s) 11.4691 +time/total (s) 8501.33 +Epoch -181 +---------------------------------- --------------- +2022-05-10 15:32:39.758928 PDT | [1] Epoch -180 finished +---------------------------------- --------------- +epoch -180 +replay_buffer/size 999033 +trainer/num train calls 821000 +trainer/Policy Loss -20.0293 +trainer/Log Pis Mean 24.662 +trainer/Log Pis Std 13.4018 +trainer/Log Pis Max 65.0969 +trainer/Log Pis Min -9.99092 +trainer/policy/mean Mean -0.0384468 +trainer/policy/mean Std 0.909043 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77804 +trainer/policy/normal/std Std 0.651335 +trainer/policy/normal/std Max 5.21365 +trainer/policy/normal/std Min 0.232904 +trainer/policy/normal/log_std Mean 0.983156 +trainer/policy/normal/log_std Std 0.31105 +trainer/policy/normal/log_std Max 1.65128 +trainer/policy/normal/log_std Min -1.45713 +eval/num steps total 818715 +eval/num paths total 821 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.335118 +eval/Actions Std 0.887396 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60209 +time/logging (s) 0.00425174 +time/sampling batch (s) 0.276153 +time/saving (s) 0.00389246 +time/training (s) 6.42554 +time/epoch (s) 9.31193 +time/total (s) 8510.65 +Epoch -180 +---------------------------------- --------------- +2022-05-10 15:32:50.225107 PDT | [1] Epoch -179 finished +---------------------------------- --------------- +epoch -179 +replay_buffer/size 999033 +trainer/num train calls 822000 +trainer/Policy Loss -18.3517 +trainer/Log Pis Mean 23.8405 +trainer/Log Pis Std 13.8788 +trainer/Log Pis Max 68.4131 +trainer/Log Pis Min -12.2404 +trainer/policy/mean Mean -0.0363717 +trainer/policy/mean Std 0.907672 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84492 +trainer/policy/normal/std Std 0.658164 +trainer/policy/normal/std Max 5.5746 +trainer/policy/normal/std Min 0.303918 +trainer/policy/normal/log_std Mean 1.00831 +trainer/policy/normal/log_std Std 0.305344 +trainer/policy/normal/log_std Max 1.71822 +trainer/policy/normal/log_std Min -1.191 +eval/num steps total 819715 +eval/num paths total 822 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.179966 +eval/Actions Std 0.859532 +eval/Actions Max 0.999993 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75215 +time/logging (s) 0.00390045 +time/sampling batch (s) 0.779827 +time/saving (s) 0.00371539 +time/training (s) 6.90102 +time/epoch (s) 10.4406 +time/total (s) 8521.09 +Epoch -179 +---------------------------------- --------------- +2022-05-10 15:33:00.841181 PDT | [1] Epoch -178 finished +---------------------------------- --------------- +epoch -178 +replay_buffer/size 999033 +trainer/num train calls 823000 +trainer/Policy Loss -19.6889 +trainer/Log Pis Mean 23.7405 +trainer/Log Pis Std 12.9385 +trainer/Log Pis Max 62.5731 +trainer/Log Pis Min -11.2921 +trainer/policy/mean Mean -0.030074 +trainer/policy/mean Std 0.906438 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8207 +trainer/policy/normal/std Std 0.676665 +trainer/policy/normal/std Max 5.59313 +trainer/policy/normal/std Min 0.32537 +trainer/policy/normal/log_std Mean 0.994042 +trainer/policy/normal/log_std Std 0.333683 +trainer/policy/normal/log_std Max 1.72154 +trainer/policy/normal/log_std Min -1.12279 +eval/num steps total 820715 +eval/num paths total 823 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0215934 +eval/Actions Std 0.911881 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39222 +time/logging (s) 0.0037971 +time/sampling batch (s) 0.280893 +time/saving (s) 0.00347028 +time/training (s) 7.91067 +time/epoch (s) 10.591 +time/total (s) 8531.68 +Epoch -178 +---------------------------------- --------------- +2022-05-10 15:33:09.909579 PDT | [1] Epoch -177 finished +---------------------------------- -------------- +epoch -177 +replay_buffer/size 999033 +trainer/num train calls 824000 +trainer/Policy Loss -18.1862 +trainer/Log Pis Mean 23.9505 +trainer/Log Pis Std 12.7729 +trainer/Log Pis Max 66.9889 +trainer/Log Pis Min -7.84992 +trainer/policy/mean Mean -0.0282782 +trainer/policy/mean Std 0.907163 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99997 +trainer/policy/normal/std Mean 2.7844 +trainer/policy/normal/std Std 0.640189 +trainer/policy/normal/std Max 5.4924 +trainer/policy/normal/std Min 0.324882 +trainer/policy/normal/log_std Mean 0.985695 +trainer/policy/normal/log_std Std 0.312288 +trainer/policy/normal/log_std Max 1.70337 +trainer/policy/normal/log_std Min -1.12429 +eval/num steps total 821715 +eval/num paths total 824 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.173224 +eval/Actions Std 0.91319 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58295 +time/logging (s) 0.0037918 +time/sampling batch (s) 0.280037 +time/saving (s) 0.0034095 +time/training (s) 6.17355 +time/epoch (s) 9.04374 +time/total (s) 8540.73 +Epoch -177 +---------------------------------- -------------- +2022-05-10 15:33:20.286765 PDT | [1] Epoch -176 finished +---------------------------------- --------------- +epoch -176 +replay_buffer/size 999033 +trainer/num train calls 825000 +trainer/Policy Loss -19.0544 +trainer/Log Pis Mean 24.4721 +trainer/Log Pis Std 12.6324 +trainer/Log Pis Max 63.5319 +trainer/Log Pis Min -7.61943 +trainer/policy/mean Mean -0.0403195 +trainer/policy/mean Std 0.9075 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8559 +trainer/policy/normal/std Std 0.663335 +trainer/policy/normal/std Max 5.56784 +trainer/policy/normal/std Min 0.254409 +trainer/policy/normal/log_std Mean 1.01017 +trainer/policy/normal/log_std Std 0.31827 +trainer/policy/normal/log_std Max 1.71701 +trainer/policy/normal/log_std Min -1.36881 +eval/num steps total 822715 +eval/num paths total 825 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.013471 +eval/Actions Std 0.886068 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3796 +time/logging (s) 0.00375744 +time/sampling batch (s) 0.283459 +time/saving (s) 0.00347777 +time/training (s) 7.68201 +time/epoch (s) 10.3523 +time/total (s) 8551.09 +Epoch -176 +---------------------------------- --------------- +2022-05-10 15:33:29.603139 PDT | [1] Epoch -175 finished +---------------------------------- --------------- +epoch -175 +replay_buffer/size 999033 +trainer/num train calls 826000 +trainer/Policy Loss -19.6548 +trainer/Log Pis Mean 25.0681 +trainer/Log Pis Std 13.4218 +trainer/Log Pis Max 75.2991 +trainer/Log Pis Min -6.091 +trainer/policy/mean Mean -0.0438632 +trainer/policy/mean Std 0.912974 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79456 +trainer/policy/normal/std Std 0.66935 +trainer/policy/normal/std Max 6.12026 +trainer/policy/normal/std Min 0.292977 +trainer/policy/normal/log_std Mean 0.986156 +trainer/policy/normal/log_std Std 0.326075 +trainer/policy/normal/log_std Max 1.8116 +trainer/policy/normal/log_std Min -1.22766 +eval/num steps total 823715 +eval/num paths total 826 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0513044 +eval/Actions Std 0.848672 +eval/Actions Max 0.999978 +eval/Actions Min -0.999974 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49838 +time/logging (s) 0.00393151 +time/sampling batch (s) 0.782898 +time/saving (s) 0.00346671 +time/training (s) 6.00319 +time/epoch (s) 9.29187 +time/total (s) 8560.38 +Epoch -175 +---------------------------------- --------------- +2022-05-10 15:33:39.600076 PDT | [1] Epoch -174 finished +---------------------------------- --------------- +epoch -174 +replay_buffer/size 999033 +trainer/num train calls 827000 +trainer/Policy Loss -19.5076 +trainer/Log Pis Mean 24.6984 +trainer/Log Pis Std 13.6937 +trainer/Log Pis Max 73.7412 +trainer/Log Pis Min -7.07835 +trainer/policy/mean Mean -0.0309103 +trainer/policy/mean Std 0.912436 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82239 +trainer/policy/normal/std Std 0.636128 +trainer/policy/normal/std Max 6.87317 +trainer/policy/normal/std Min 0.263098 +trainer/policy/normal/log_std Mean 1.00183 +trainer/policy/normal/log_std Std 0.300003 +trainer/policy/normal/log_std Max 1.92763 +trainer/policy/normal/log_std Min -1.33523 +eval/num steps total 824715 +eval/num paths total 827 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0679768 +eval/Actions Std 0.910838 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.20622 +time/logging (s) 0.00425301 +time/sampling batch (s) 0.531115 +time/saving (s) 0.00413996 +time/training (s) 7.22672 +time/epoch (s) 9.97244 +time/total (s) 8570.36 +Epoch -174 +---------------------------------- --------------- +2022-05-10 15:33:49.653347 PDT | [1] Epoch -173 finished +---------------------------------- --------------- +epoch -173 +replay_buffer/size 999033 +trainer/num train calls 828000 +trainer/Policy Loss -20.6669 +trainer/Log Pis Mean 24.7555 +trainer/Log Pis Std 14.1253 +trainer/Log Pis Max 70.1364 +trainer/Log Pis Min -11.5933 +trainer/policy/mean Mean -0.0310479 +trainer/policy/mean Std 0.905341 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87714 +trainer/policy/normal/std Std 0.747926 +trainer/policy/normal/std Max 6.20006 +trainer/policy/normal/std Min 0.230846 +trainer/policy/normal/log_std Mean 1.00602 +trainer/policy/normal/log_std Std 0.365609 +trainer/policy/normal/log_std Max 1.82456 +trainer/policy/normal/log_std Min -1.466 +eval/num steps total 825715 +eval/num paths total 828 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0185788 +eval/Actions Std 0.915588 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67018 +time/logging (s) 0.00382845 +time/sampling batch (s) 0.279508 +time/saving (s) 0.00356317 +time/training (s) 7.0707 +time/epoch (s) 10.0278 +time/total (s) 8580.39 +Epoch -173 +---------------------------------- --------------- +2022-05-10 15:33:59.218450 PDT | [1] Epoch -172 finished +---------------------------------- --------------- +epoch -172 +replay_buffer/size 999033 +trainer/num train calls 829000 +trainer/Policy Loss -19.4208 +trainer/Log Pis Mean 23.8952 +trainer/Log Pis Std 12.956 +trainer/Log Pis Max 62.2092 +trainer/Log Pis Min -8.19847 +trainer/policy/mean Mean -0.0300232 +trainer/policy/mean Std 0.904275 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82781 +trainer/policy/normal/std Std 0.697264 +trainer/policy/normal/std Max 5.60333 +trainer/policy/normal/std Min 0.21706 +trainer/policy/normal/log_std Mean 0.992988 +trainer/policy/normal/log_std Std 0.350652 +trainer/policy/normal/log_std Max 1.72336 +trainer/policy/normal/log_std Min -1.52758 +eval/num steps total 826715 +eval/num paths total 829 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0354815 +eval/Actions Std 0.908992 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56626 +time/logging (s) 0.00384634 +time/sampling batch (s) 0.27911 +time/saving (s) 0.0035112 +time/training (s) 6.6876 +time/epoch (s) 9.54033 +time/total (s) 8589.93 +Epoch -172 +---------------------------------- --------------- +2022-05-10 15:34:08.715011 PDT | [1] Epoch -171 finished +---------------------------------- --------------- +epoch -171 +replay_buffer/size 999033 +trainer/num train calls 830000 +trainer/Policy Loss -20.0819 +trainer/Log Pis Mean 23.7551 +trainer/Log Pis Std 12.5981 +trainer/Log Pis Max 58.0375 +trainer/Log Pis Min -10.91 +trainer/policy/mean Mean -0.0431239 +trainer/policy/mean Std 0.90757 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.74758 +trainer/policy/normal/std Std 0.666429 +trainer/policy/normal/std Max 5.40979 +trainer/policy/normal/std Min 0.240401 +trainer/policy/normal/log_std Mean 0.966378 +trainer/policy/normal/log_std Std 0.33996 +trainer/policy/normal/log_std Max 1.68821 +trainer/policy/normal/log_std Min -1.42545 +eval/num steps total 827715 +eval/num paths total 830 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141224 +eval/Actions Std 0.921087 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57749 +time/logging (s) 0.00373181 +time/sampling batch (s) 0.284338 +time/saving (s) 0.00342843 +time/training (s) 6.60263 +time/epoch (s) 9.47162 +time/total (s) 8599.41 +Epoch -171 +---------------------------------- --------------- +2022-05-10 15:34:20.013144 PDT | [1] Epoch -170 finished +---------------------------------- --------------- +epoch -170 +replay_buffer/size 999033 +trainer/num train calls 831000 +trainer/Policy Loss -18.9431 +trainer/Log Pis Mean 24.6149 +trainer/Log Pis Std 13.2218 +trainer/Log Pis Max 61.6931 +trainer/Log Pis Min -2.99796 +trainer/policy/mean Mean -0.0316069 +trainer/policy/mean Std 0.907442 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80511 +trainer/policy/normal/std Std 0.681004 +trainer/policy/normal/std Max 5.67475 +trainer/policy/normal/std Min 0.212019 +trainer/policy/normal/log_std Mean 0.987186 +trainer/policy/normal/log_std Std 0.339387 +trainer/policy/normal/log_std Max 1.73603 +trainer/policy/normal/log_std Min -1.55108 +eval/num steps total 828715 +eval/num paths total 831 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109956 +eval/Actions Std 0.917358 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58378 +time/logging (s) 0.00390007 +time/sampling batch (s) 0.529222 +time/saving (s) 0.00349024 +time/training (s) 8.15339 +time/epoch (s) 11.2738 +time/total (s) 8610.68 +Epoch -170 +---------------------------------- --------------- +2022-05-10 15:34:30.896683 PDT | [1] Epoch -169 finished +---------------------------------- --------------- +epoch -169 +replay_buffer/size 999033 +trainer/num train calls 832000 +trainer/Policy Loss -18.9397 +trainer/Log Pis Mean 24.4656 +trainer/Log Pis Std 13.3811 +trainer/Log Pis Max 62.7137 +trainer/Log Pis Min -13.0514 +trainer/policy/mean Mean -0.0635042 +trainer/policy/mean Std 0.905341 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78433 +trainer/policy/normal/std Std 0.656642 +trainer/policy/normal/std Max 6.09 +trainer/policy/normal/std Min 0.268124 +trainer/policy/normal/log_std Mean 0.982648 +trainer/policy/normal/log_std Std 0.327353 +trainer/policy/normal/log_std Max 1.80665 +trainer/policy/normal/log_std Min -1.31631 +eval/num steps total 829715 +eval/num paths total 832 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0206834 +eval/Actions Std 0.91154 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65541 +time/logging (s) 0.00415621 +time/sampling batch (s) 0.279001 +time/saving (s) 0.00382431 +time/training (s) 7.91669 +time/epoch (s) 10.8591 +time/total (s) 8621.54 +Epoch -169 +---------------------------------- --------------- +2022-05-10 15:34:41.304797 PDT | [1] Epoch -168 finished +---------------------------------- --------------- +epoch -168 +replay_buffer/size 999033 +trainer/num train calls 833000 +trainer/Policy Loss -19.5135 +trainer/Log Pis Mean 24.9977 +trainer/Log Pis Std 13.7138 +trainer/Log Pis Max 67.7228 +trainer/Log Pis Min -7.02795 +trainer/policy/mean Mean -0.0384229 +trainer/policy/mean Std 0.909571 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75601 +trainer/policy/normal/std Std 0.661392 +trainer/policy/normal/std Max 6.30243 +trainer/policy/normal/std Min 0.215481 +trainer/policy/normal/log_std Mean 0.970896 +trainer/policy/normal/log_std Std 0.334549 +trainer/policy/normal/log_std Max 1.84094 +trainer/policy/normal/log_std Min -1.53488 +eval/num steps total 830715 +eval/num paths total 833 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.159105 +eval/Actions Std 0.859396 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73819 +time/logging (s) 0.00392895 +time/sampling batch (s) 0.280661 +time/saving (s) 0.00376657 +time/training (s) 7.35608 +time/epoch (s) 10.3826 +time/total (s) 8631.93 +Epoch -168 +---------------------------------- --------------- +2022-05-10 15:34:50.763398 PDT | [1] Epoch -167 finished +---------------------------------- --------------- +epoch -167 +replay_buffer/size 999033 +trainer/num train calls 834000 +trainer/Policy Loss -19.7057 +trainer/Log Pis Mean 24.6521 +trainer/Log Pis Std 13.317 +trainer/Log Pis Max 70.6564 +trainer/Log Pis Min -9.82476 +trainer/policy/mean Mean -0.0102961 +trainer/policy/mean Std 0.909148 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.75421 +trainer/policy/normal/std Std 0.6599 +trainer/policy/normal/std Max 7.02138 +trainer/policy/normal/std Min 0.282351 +trainer/policy/normal/log_std Mean 0.971492 +trainer/policy/normal/log_std Std 0.325275 +trainer/policy/normal/log_std Max 1.94896 +trainer/policy/normal/log_std Min -1.2646 +eval/num steps total 831715 +eval/num paths total 834 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0468424 +eval/Actions Std 0.91343 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28298 +time/logging (s) 0.00376006 +time/sampling batch (s) 0.2786 +time/saving (s) 0.00345423 +time/training (s) 6.86481 +time/epoch (s) 9.4336 +time/total (s) 8641.37 +Epoch -167 +---------------------------------- --------------- +2022-05-10 15:35:01.398254 PDT | [1] Epoch -166 finished +---------------------------------- --------------- +epoch -166 +replay_buffer/size 999033 +trainer/num train calls 835000 +trainer/Policy Loss -19.6814 +trainer/Log Pis Mean 23.8309 +trainer/Log Pis Std 12.9372 +trainer/Log Pis Max 65.5446 +trainer/Log Pis Min -9.3877 +trainer/policy/mean Mean -0.0506658 +trainer/policy/mean Std 0.903764 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.75043 +trainer/policy/normal/std Std 0.66491 +trainer/policy/normal/std Max 6.95365 +trainer/policy/normal/std Min 0.241413 +trainer/policy/normal/log_std Mean 0.967787 +trainer/policy/normal/log_std Std 0.339596 +trainer/policy/normal/log_std Max 1.93927 +trainer/policy/normal/log_std Min -1.42124 +eval/num steps total 832715 +eval/num paths total 835 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.27633 +eval/Actions Std 0.878156 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56109 +time/logging (s) 0.00375551 +time/sampling batch (s) 0.528558 +time/saving (s) 0.00345875 +time/training (s) 7.51328 +time/epoch (s) 10.6102 +time/total (s) 8651.98 +Epoch -166 +---------------------------------- --------------- +2022-05-10 15:35:11.909352 PDT | [1] Epoch -165 finished +---------------------------------- --------------- +epoch -165 +replay_buffer/size 999033 +trainer/num train calls 836000 +trainer/Policy Loss -20.6374 +trainer/Log Pis Mean 24.4371 +trainer/Log Pis Std 13.7882 +trainer/Log Pis Max 64.8963 +trainer/Log Pis Min -8.65842 +trainer/policy/mean Mean -0.0449172 +trainer/policy/mean Std 0.913237 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.75784 +trainer/policy/normal/std Std 0.674901 +trainer/policy/normal/std Max 6.59086 +trainer/policy/normal/std Min 0.262224 +trainer/policy/normal/log_std Mean 0.970147 +trainer/policy/normal/log_std Std 0.338329 +trainer/policy/normal/log_std Max 1.88568 +trainer/policy/normal/log_std Min -1.33855 +eval/num steps total 833715 +eval/num paths total 836 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0987991 +eval/Actions Std 0.899667 +eval/Actions Max 0.999979 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64462 +time/logging (s) 0.00377937 +time/sampling batch (s) 0.287568 +time/saving (s) 0.00368892 +time/training (s) 7.54686 +time/epoch (s) 10.4865 +time/total (s) 8662.47 +Epoch -165 +---------------------------------- --------------- +2022-05-10 15:35:22.737751 PDT | [1] Epoch -164 finished +---------------------------------- --------------- +epoch -164 +replay_buffer/size 999033 +trainer/num train calls 837000 +trainer/Policy Loss -19.2395 +trainer/Log Pis Mean 23.2237 +trainer/Log Pis Std 12.9904 +trainer/Log Pis Max 72.4765 +trainer/Log Pis Min -7.7688 +trainer/policy/mean Mean -0.0208968 +trainer/policy/mean Std 0.907355 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.73538 +trainer/policy/normal/std Std 0.646494 +trainer/policy/normal/std Max 5.93509 +trainer/policy/normal/std Min 0.277575 +trainer/policy/normal/log_std Mean 0.967908 +trainer/policy/normal/log_std Std 0.309181 +trainer/policy/normal/log_std Max 1.78088 +trainer/policy/normal/log_std Min -1.28167 +eval/num steps total 834715 +eval/num paths total 837 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0206555 +eval/Actions Std 0.909384 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67575 +time/logging (s) 0.00412021 +time/sampling batch (s) 0.292342 +time/saving (s) 0.00382782 +time/training (s) 7.8277 +time/epoch (s) 10.8037 +time/total (s) 8673.28 +Epoch -164 +---------------------------------- --------------- +2022-05-10 15:35:33.864774 PDT | [1] Epoch -163 finished +---------------------------------- --------------- +epoch -163 +replay_buffer/size 999033 +trainer/num train calls 838000 +trainer/Policy Loss -18.384 +trainer/Log Pis Mean 25.3136 +trainer/Log Pis Std 13.1339 +trainer/Log Pis Max 63.4259 +trainer/Log Pis Min -10.6278 +trainer/policy/mean Mean -0.0340906 +trainer/policy/mean Std 0.907142 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76999 +trainer/policy/normal/std Std 0.668936 +trainer/policy/normal/std Max 5.43541 +trainer/policy/normal/std Min 0.233123 +trainer/policy/normal/log_std Mean 0.974697 +trainer/policy/normal/log_std Std 0.340576 +trainer/policy/normal/log_std Max 1.69294 +trainer/policy/normal/log_std Min -1.45619 +eval/num steps total 835715 +eval/num paths total 838 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.222481 +eval/Actions Std 0.909351 +eval/Actions Max 0.999975 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61647 +time/logging (s) 0.00390302 +time/sampling batch (s) 0.780073 +time/saving (s) 0.00370681 +time/training (s) 7.69748 +time/epoch (s) 11.1016 +time/total (s) 8684.38 +Epoch -163 +---------------------------------- --------------- +2022-05-10 15:35:44.054441 PDT | [1] Epoch -162 finished +---------------------------------- --------------- +epoch -162 +replay_buffer/size 999033 +trainer/num train calls 839000 +trainer/Policy Loss -19.2827 +trainer/Log Pis Mean 24.8971 +trainer/Log Pis Std 13.59 +trainer/Log Pis Max 78.5778 +trainer/Log Pis Min -4.13524 +trainer/policy/mean Mean -0.0465017 +trainer/policy/mean Std 0.907873 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80585 +trainer/policy/normal/std Std 0.674679 +trainer/policy/normal/std Max 5.90254 +trainer/policy/normal/std Min 0.248534 +trainer/policy/normal/log_std Mean 0.990275 +trainer/policy/normal/log_std Std 0.324832 +trainer/policy/normal/log_std Max 1.77538 +trainer/policy/normal/log_std Min -1.39218 +eval/num steps total 836715 +eval/num paths total 839 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.247188 +eval/Actions Std 0.866365 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65276 +time/logging (s) 0.00374268 +time/sampling batch (s) 0.281974 +time/saving (s) 0.00344375 +time/training (s) 7.2227 +time/epoch (s) 10.1646 +time/total (s) 8694.55 +Epoch -162 +---------------------------------- --------------- +2022-05-10 15:35:54.030319 PDT | [1] Epoch -161 finished +---------------------------------- --------------- +epoch -161 +replay_buffer/size 999033 +trainer/num train calls 840000 +trainer/Policy Loss -19.5633 +trainer/Log Pis Mean 24.5314 +trainer/Log Pis Std 13.5823 +trainer/Log Pis Max 72.6202 +trainer/Log Pis Min -8.47762 +trainer/policy/mean Mean -0.041619 +trainer/policy/mean Std 0.905821 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79733 +trainer/policy/normal/std Std 0.66503 +trainer/policy/normal/std Max 5.5927 +trainer/policy/normal/std Min 0.324548 +trainer/policy/normal/log_std Mean 0.987074 +trainer/policy/normal/log_std Std 0.328258 +trainer/policy/normal/log_std Max 1.72146 +trainer/policy/normal/log_std Min -1.12532 +eval/num steps total 837715 +eval/num paths total 840 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0878851 +eval/Actions Std 0.878243 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72707 +time/logging (s) 0.00378457 +time/sampling batch (s) 0.280884 +time/saving (s) 0.00346728 +time/training (s) 6.93595 +time/epoch (s) 9.95115 +time/total (s) 8704.5 +Epoch -161 +---------------------------------- --------------- +2022-05-10 15:36:04.245396 PDT | [1] Epoch -160 finished +---------------------------------- --------------- +epoch -160 +replay_buffer/size 999033 +trainer/num train calls 841000 +trainer/Policy Loss -19.6295 +trainer/Log Pis Mean 24.5113 +trainer/Log Pis Std 13.4014 +trainer/Log Pis Max 68.0466 +trainer/Log Pis Min -10.7176 +trainer/policy/mean Mean -0.0640239 +trainer/policy/mean Std 0.909071 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81827 +trainer/policy/normal/std Std 0.665366 +trainer/policy/normal/std Max 6.17406 +trainer/policy/normal/std Min 0.306536 +trainer/policy/normal/log_std Mean 0.995851 +trainer/policy/normal/log_std Std 0.320389 +trainer/policy/normal/log_std Max 1.82036 +trainer/policy/normal/log_std Min -1.18242 +eval/num steps total 838715 +eval/num paths total 841 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.157083 +eval/Actions Std 0.859875 +eval/Actions Max 0.999989 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7473 +time/logging (s) 0.00377533 +time/sampling batch (s) 0.279546 +time/saving (s) 0.0035225 +time/training (s) 7.15633 +time/epoch (s) 10.1905 +time/total (s) 8714.7 +Epoch -160 +---------------------------------- --------------- +2022-05-10 15:36:14.319223 PDT | [1] Epoch -159 finished +---------------------------------- --------------- +epoch -159 +replay_buffer/size 999033 +trainer/num train calls 842000 +trainer/Policy Loss -20.634 +trainer/Log Pis Mean 25.036 +trainer/Log Pis Std 14.0638 +trainer/Log Pis Max 73.2114 +trainer/Log Pis Min -5.20031 +trainer/policy/mean Mean -0.0359639 +trainer/policy/mean Std 0.905887 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.78987 +trainer/policy/normal/std Std 0.683386 +trainer/policy/normal/std Max 6.77081 +trainer/policy/normal/std Min 0.272923 +trainer/policy/normal/log_std Mean 0.981586 +trainer/policy/normal/log_std Std 0.340025 +trainer/policy/normal/log_std Max 1.91262 +trainer/policy/normal/log_std Min -1.29856 +eval/num steps total 839446 +eval/num paths total 842 +eval/path length Mean 731 +eval/path length Std 0 +eval/path length Max 731 +eval/path length Min 731 +eval/Rewards Mean 0.00136799 +eval/Rewards Std 0.036961 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.00714619 +eval/Actions Std 0.90396 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.59179 +time/logging (s) 0.0030206 +time/sampling batch (s) 0.525079 +time/saving (s) 0.00340398 +time/training (s) 6.92527 +time/epoch (s) 10.0486 +time/total (s) 8724.75 +Epoch -159 +---------------------------------- --------------- +2022-05-10 15:36:24.615171 PDT | [1] Epoch -158 finished +---------------------------------- --------------- +epoch -158 +replay_buffer/size 999033 +trainer/num train calls 843000 +trainer/Policy Loss -20.9074 +trainer/Log Pis Mean 24.1078 +trainer/Log Pis Std 13.3343 +trainer/Log Pis Max 73.2899 +trainer/Log Pis Min -7.45386 +trainer/policy/mean Mean -0.0213379 +trainer/policy/mean Std 0.911191 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83898 +trainer/policy/normal/std Std 0.67734 +trainer/policy/normal/std Max 7.00667 +trainer/policy/normal/std Min 0.280372 +trainer/policy/normal/log_std Mean 1.00353 +trainer/policy/normal/log_std Std 0.318047 +trainer/policy/normal/log_std Max 1.94686 +trainer/policy/normal/log_std Min -1.27164 +eval/num steps total 840446 +eval/num paths total 843 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0683217 +eval/Actions Std 0.917149 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56228 +time/logging (s) 0.00408605 +time/sampling batch (s) 0.525278 +time/saving (s) 0.00385082 +time/training (s) 7.17708 +time/epoch (s) 10.2726 +time/total (s) 8735.02 +Epoch -158 +---------------------------------- --------------- +2022-05-10 15:36:35.447828 PDT | [1] Epoch -157 finished +---------------------------------- --------------- +epoch -157 +replay_buffer/size 999033 +trainer/num train calls 844000 +trainer/Policy Loss -20.534 +trainer/Log Pis Mean 23.0303 +trainer/Log Pis Std 12.523 +trainer/Log Pis Max 60.3064 +trainer/Log Pis Min -7.83027 +trainer/policy/mean Mean -0.0375783 +trainer/policy/mean Std 0.90439 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77175 +trainer/policy/normal/std Std 0.693278 +trainer/policy/normal/std Max 6.03739 +trainer/policy/normal/std Min 0.246537 +trainer/policy/normal/log_std Mean 0.973054 +trainer/policy/normal/log_std Std 0.347628 +trainer/policy/normal/log_std Max 1.79797 +trainer/policy/normal/log_std Min -1.40024 +eval/num steps total 841446 +eval/num paths total 844 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.03021 +eval/Actions Std 0.901884 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72483 +time/logging (s) 0.00373522 +time/sampling batch (s) 0.52724 +time/saving (s) 0.00353257 +time/training (s) 7.54816 +time/epoch (s) 10.8075 +time/total (s) 8745.83 +Epoch -157 +---------------------------------- --------------- +2022-05-10 15:36:45.191006 PDT | [1] Epoch -156 finished +---------------------------------- --------------- +epoch -156 +replay_buffer/size 999033 +trainer/num train calls 845000 +trainer/Policy Loss -20.3447 +trainer/Log Pis Mean 25.1819 +trainer/Log Pis Std 12.9211 +trainer/Log Pis Max 67.9435 +trainer/Log Pis Min -7.06437 +trainer/policy/mean Mean -0.0246939 +trainer/policy/mean Std 0.911435 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.7676 +trainer/policy/normal/std Std 0.681983 +trainer/policy/normal/std Max 5.37621 +trainer/policy/normal/std Min 0.227395 +trainer/policy/normal/log_std Mean 0.972106 +trainer/policy/normal/log_std Std 0.346778 +trainer/policy/normal/log_std Max 1.68198 +trainer/policy/normal/log_std Min -1.48107 +eval/num steps total 842446 +eval/num paths total 845 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0150064 +eval/Actions Std 0.905672 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63421 +time/logging (s) 0.00391755 +time/sampling batch (s) 0.278459 +time/saving (s) 0.00356961 +time/training (s) 6.79839 +time/epoch (s) 9.71855 +time/total (s) 8755.56 +Epoch -156 +---------------------------------- --------------- +2022-05-10 15:36:54.957264 PDT | [1] Epoch -155 finished +---------------------------------- --------------- +epoch -155 +replay_buffer/size 999033 +trainer/num train calls 846000 +trainer/Policy Loss -20.1928 +trainer/Log Pis Mean 25.1165 +trainer/Log Pis Std 13.2246 +trainer/Log Pis Max 69.779 +trainer/Log Pis Min -8.40375 +trainer/policy/mean Mean -0.0730911 +trainer/policy/mean Std 0.909033 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7989 +trainer/policy/normal/std Std 0.637112 +trainer/policy/normal/std Max 5.47593 +trainer/policy/normal/std Min 0.329099 +trainer/policy/normal/log_std Mean 0.991803 +trainer/policy/normal/log_std Std 0.308653 +trainer/policy/normal/log_std Max 1.70036 +trainer/policy/normal/log_std Min -1.1114 +eval/num steps total 843446 +eval/num paths total 846 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0937183 +eval/Actions Std 0.916619 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6418 +time/logging (s) 0.00375159 +time/sampling batch (s) 0.779726 +time/saving (s) 0.00344564 +time/training (s) 6.31232 +time/epoch (s) 9.74104 +time/total (s) 8765.3 +Epoch -155 +---------------------------------- --------------- +2022-05-10 15:37:06.469669 PDT | [1] Epoch -154 finished +---------------------------------- --------------- +epoch -154 +replay_buffer/size 999033 +trainer/num train calls 847000 +trainer/Policy Loss -20.4131 +trainer/Log Pis Mean 24.1641 +trainer/Log Pis Std 12.6131 +trainer/Log Pis Max 62.936 +trainer/Log Pis Min -11.8971 +trainer/policy/mean Mean -0.0473196 +trainer/policy/mean Std 0.908764 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75359 +trainer/policy/normal/std Std 0.700793 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.25405 +trainer/policy/normal/log_std Mean 0.966318 +trainer/policy/normal/log_std Std 0.345584 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.37022 +eval/num steps total 844446 +eval/num paths total 847 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.41604 +eval/Actions Std 0.819744 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7398 +time/logging (s) 0.00429793 +time/sampling batch (s) 1.28022 +time/saving (s) 0.00413099 +time/training (s) 7.4598 +time/epoch (s) 11.4883 +time/total (s) 8776.79 +Epoch -154 +---------------------------------- --------------- +2022-05-10 15:37:15.329099 PDT | [1] Epoch -153 finished +---------------------------------- --------------- +epoch -153 +replay_buffer/size 999033 +trainer/num train calls 848000 +trainer/Policy Loss -21.4743 +trainer/Log Pis Mean 24.4479 +trainer/Log Pis Std 12.9604 +trainer/Log Pis Max 61.0412 +trainer/Log Pis Min -9.44162 +trainer/policy/mean Mean -0.0615296 +trainer/policy/mean Std 0.907696 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77733 +trainer/policy/normal/std Std 0.642279 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.298706 +trainer/policy/normal/log_std Mean 0.983601 +trainer/policy/normal/log_std Std 0.310255 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20829 +eval/num steps total 845446 +eval/num paths total 848 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.209437 +eval/Actions Std 0.888313 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51561 +time/logging (s) 0.00372008 +time/sampling batch (s) 0.274283 +time/saving (s) 0.0034598 +time/training (s) 6.03673 +time/epoch (s) 8.8338 +time/total (s) 8785.63 +Epoch -153 +---------------------------------- --------------- +2022-05-10 15:37:25.046775 PDT | [1] Epoch -152 finished +---------------------------------- --------------- +epoch -152 +replay_buffer/size 999033 +trainer/num train calls 849000 +trainer/Policy Loss -19.3939 +trainer/Log Pis Mean 23.8069 +trainer/Log Pis Std 13.4039 +trainer/Log Pis Max 61.9076 +trainer/Log Pis Min -8.17505 +trainer/policy/mean Mean -0.0240783 +trainer/policy/mean Std 0.904438 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74555 +trainer/policy/normal/std Std 0.669069 +trainer/policy/normal/std Max 5.34151 +trainer/policy/normal/std Min 0.300372 +trainer/policy/normal/log_std Mean 0.965366 +trainer/policy/normal/log_std Std 0.340358 +trainer/policy/normal/log_std Max 1.67551 +trainer/policy/normal/log_std Min -1.20273 +eval/num steps total 846446 +eval/num paths total 849 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.466804 +eval/Actions Std 0.76449 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51571 +time/logging (s) 0.00369347 +time/sampling batch (s) 0.27731 +time/saving (s) 0.00348314 +time/training (s) 6.8929 +time/epoch (s) 9.6931 +time/total (s) 8795.33 +Epoch -152 +---------------------------------- --------------- +2022-05-10 15:37:35.596479 PDT | [1] Epoch -151 finished +---------------------------------- --------------- +epoch -151 +replay_buffer/size 999033 +trainer/num train calls 850000 +trainer/Policy Loss -19.5931 +trainer/Log Pis Mean 24.3458 +trainer/Log Pis Std 13.0651 +trainer/Log Pis Max 73.8966 +trainer/Log Pis Min -3.87758 +trainer/policy/mean Mean -0.0308535 +trainer/policy/mean Std 0.906773 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82183 +trainer/policy/normal/std Std 0.702971 +trainer/policy/normal/std Max 5.17177 +trainer/policy/normal/std Min 0.293532 +trainer/policy/normal/log_std Mean 0.990373 +trainer/policy/normal/log_std Std 0.351215 +trainer/policy/normal/log_std Max 1.64321 +trainer/policy/normal/log_std Min -1.22577 +eval/num steps total 847446 +eval/num paths total 850 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.371137 +eval/Actions Std 0.829875 +eval/Actions Max 0.999974 +eval/Actions Min -0.999961 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76184 +time/logging (s) 0.00441709 +time/sampling batch (s) 0.582737 +time/saving (s) 0.00457763 +time/training (s) 7.1712 +time/epoch (s) 10.5248 +time/total (s) 8805.85 +Epoch -151 +---------------------------------- --------------- +2022-05-10 15:37:46.289343 PDT | [1] Epoch -150 finished +---------------------------------- --------------- +epoch -150 +replay_buffer/size 999033 +trainer/num train calls 851000 +trainer/Policy Loss -19.252 +trainer/Log Pis Mean 24.8005 +trainer/Log Pis Std 13.3767 +trainer/Log Pis Max 69.0197 +trainer/Log Pis Min -5.78909 +trainer/policy/mean Mean -0.043905 +trainer/policy/mean Std 0.904524 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78308 +trainer/policy/normal/std Std 0.664512 +trainer/policy/normal/std Max 6.15684 +trainer/policy/normal/std Min 0.256351 +trainer/policy/normal/log_std Mean 0.981217 +trainer/policy/normal/log_std Std 0.332093 +trainer/policy/normal/log_std Max 1.81756 +trainer/policy/normal/log_std Min -1.36121 +eval/num steps total 848446 +eval/num paths total 851 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0108197 +eval/Actions Std 0.907869 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8044 +time/logging (s) 0.003845 +time/sampling batch (s) 0.58621 +time/saving (s) 0.00379164 +time/training (s) 7.26731 +time/epoch (s) 10.6656 +time/total (s) 8816.52 +Epoch -150 +---------------------------------- --------------- +2022-05-10 15:37:57.743562 PDT | [1] Epoch -149 finished +---------------------------------- --------------- +epoch -149 +replay_buffer/size 999033 +trainer/num train calls 852000 +trainer/Policy Loss -19.4593 +trainer/Log Pis Mean 24.0759 +trainer/Log Pis Std 13.0384 +trainer/Log Pis Max 74.2182 +trainer/Log Pis Min -3.16103 +trainer/policy/mean Mean -0.0247053 +trainer/policy/mean Std 0.903032 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75936 +trainer/policy/normal/std Std 0.660336 +trainer/policy/normal/std Max 6.08345 +trainer/policy/normal/std Min 0.200242 +trainer/policy/normal/log_std Mean 0.973612 +trainer/policy/normal/log_std Std 0.326207 +trainer/policy/normal/log_std Max 1.80557 +trainer/policy/normal/log_std Min -1.60823 +eval/num steps total 849446 +eval/num paths total 852 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00166577 +eval/Actions Std 0.90313 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53463 +time/logging (s) 0.00384765 +time/sampling batch (s) 0.579529 +time/saving (s) 0.00365101 +time/training (s) 8.30728 +time/epoch (s) 11.4289 +time/total (s) 8827.95 +Epoch -149 +---------------------------------- --------------- +2022-05-10 15:38:09.165231 PDT | [1] Epoch -148 finished +---------------------------------- --------------- +epoch -148 +replay_buffer/size 999033 +trainer/num train calls 853000 +trainer/Policy Loss -19.6595 +trainer/Log Pis Mean 24.9121 +trainer/Log Pis Std 13.3067 +trainer/Log Pis Max 68.1287 +trainer/Log Pis Min -11.8485 +trainer/policy/mean Mean -0.0514121 +trainer/policy/mean Std 0.905376 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.77557 +trainer/policy/normal/std Std 0.683026 +trainer/policy/normal/std Max 5.70393 +trainer/policy/normal/std Min 0.277445 +trainer/policy/normal/log_std Mean 0.976672 +trainer/policy/normal/log_std Std 0.337636 +trainer/policy/normal/log_std Max 1.74116 +trainer/policy/normal/log_std Min -1.28213 +eval/num steps total 850446 +eval/num paths total 853 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.186502 +eval/Actions Std 0.908208 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78739 +time/logging (s) 0.00376492 +time/sampling batch (s) 1.08078 +time/saving (s) 0.00366173 +time/training (s) 7.52064 +time/epoch (s) 11.3962 +time/total (s) 8839.35 +Epoch -148 +---------------------------------- --------------- +2022-05-10 15:38:19.964644 PDT | [1] Epoch -147 finished +---------------------------------- --------------- +epoch -147 +replay_buffer/size 999033 +trainer/num train calls 854000 +trainer/Policy Loss -20.0117 +trainer/Log Pis Mean 24.9615 +trainer/Log Pis Std 13.5942 +trainer/Log Pis Max 67.1812 +trainer/Log Pis Min -14.055 +trainer/policy/mean Mean -0.0115066 +trainer/policy/mean Std 0.907243 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.7996 +trainer/policy/normal/std Std 0.687908 +trainer/policy/normal/std Max 6.40767 +trainer/policy/normal/std Min 0.303861 +trainer/policy/normal/log_std Mean 0.986304 +trainer/policy/normal/log_std Std 0.331415 +trainer/policy/normal/log_std Max 1.8575 +trainer/policy/normal/log_std Min -1.19118 +eval/num steps total 851446 +eval/num paths total 854 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.159753 +eval/Actions Std 0.878306 +eval/Actions Max 0.999984 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82715 +time/logging (s) 0.00419648 +time/sampling batch (s) 0.328727 +time/saving (s) 0.00421734 +time/training (s) 7.61029 +time/epoch (s) 10.7746 +time/total (s) 8850.13 +Epoch -147 +---------------------------------- --------------- +2022-05-10 15:38:30.895053 PDT | [1] Epoch -146 finished +---------------------------------- --------------- +epoch -146 +replay_buffer/size 999033 +trainer/num train calls 855000 +trainer/Policy Loss -18.7407 +trainer/Log Pis Mean 24.2474 +trainer/Log Pis Std 13.302 +trainer/Log Pis Max 74.3044 +trainer/Log Pis Min -7.72905 +trainer/policy/mean Mean -0.0371761 +trainer/policy/mean Std 0.903359 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78045 +trainer/policy/normal/std Std 0.672743 +trainer/policy/normal/std Max 5.73765 +trainer/policy/normal/std Min 0.286153 +trainer/policy/normal/log_std Mean 0.979537 +trainer/policy/normal/log_std Std 0.334007 +trainer/policy/normal/log_std Max 1.74705 +trainer/policy/normal/log_std Min -1.25123 +eval/num steps total 852446 +eval/num paths total 855 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.150154 +eval/Actions Std 0.859861 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.94674 +time/logging (s) 0.00421166 +time/sampling batch (s) 1.05204 +time/saving (s) 0.00347391 +time/training (s) 6.89826 +time/epoch (s) 10.9047 +time/total (s) 8861.04 +Epoch -146 +---------------------------------- --------------- +2022-05-10 15:38:40.489456 PDT | [1] Epoch -145 finished +---------------------------------- --------------- +epoch -145 +replay_buffer/size 999033 +trainer/num train calls 856000 +trainer/Policy Loss -19.762 +trainer/Log Pis Mean 23.8847 +trainer/Log Pis Std 12.9431 +trainer/Log Pis Max 70.1287 +trainer/Log Pis Min -6.26433 +trainer/policy/mean Mean -0.0263758 +trainer/policy/mean Std 0.906305 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85127 +trainer/policy/normal/std Std 0.702637 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.252513 +trainer/policy/normal/log_std Mean 1.00283 +trainer/policy/normal/log_std Std 0.342029 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.37629 +eval/num steps total 853446 +eval/num paths total 856 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.023689 +eval/Actions Std 0.895711 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44463 +time/logging (s) 0.00409605 +time/sampling batch (s) 0.278304 +time/saving (s) 0.00383677 +time/training (s) 6.83847 +time/epoch (s) 9.56934 +time/total (s) 8870.61 +Epoch -145 +---------------------------------- --------------- +2022-05-10 15:38:50.311381 PDT | [1] Epoch -144 finished +---------------------------------- --------------- +epoch -144 +replay_buffer/size 999033 +trainer/num train calls 857000 +trainer/Policy Loss -18.8379 +trainer/Log Pis Mean 25.3739 +trainer/Log Pis Std 13.0488 +trainer/Log Pis Max 71.0848 +trainer/Log Pis Min -8.60889 +trainer/policy/mean Mean -0.034596 +trainer/policy/mean Std 0.911477 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80541 +trainer/policy/normal/std Std 0.664282 +trainer/policy/normal/std Max 5.3093 +trainer/policy/normal/std Min 0.337202 +trainer/policy/normal/log_std Mean 0.991341 +trainer/policy/normal/log_std Std 0.318931 +trainer/policy/normal/log_std Max 1.66946 +trainer/policy/normal/log_std Min -1.08707 +eval/num steps total 854446 +eval/num paths total 857 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.306509 +eval/Actions Std 0.853557 +eval/Actions Max 0.999979 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6195 +time/logging (s) 0.00377624 +time/sampling batch (s) 0.524285 +time/saving (s) 0.00368729 +time/training (s) 6.64537 +time/epoch (s) 9.79662 +time/total (s) 8880.41 +Epoch -144 +---------------------------------- --------------- +2022-05-10 15:39:00.183811 PDT | [1] Epoch -143 finished +---------------------------------- --------------- +epoch -143 +replay_buffer/size 999033 +trainer/num train calls 858000 +trainer/Policy Loss -21.0726 +trainer/Log Pis Mean 26.1168 +trainer/Log Pis Std 13.1099 +trainer/Log Pis Max 76.0938 +trainer/Log Pis Min -13.2067 +trainer/policy/mean Mean -0.0311493 +trainer/policy/mean Std 0.909344 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81215 +trainer/policy/normal/std Std 0.672959 +trainer/policy/normal/std Max 5.6538 +trainer/policy/normal/std Min 0.240432 +trainer/policy/normal/log_std Mean 0.992545 +trainer/policy/normal/log_std Std 0.325648 +trainer/policy/normal/log_std Max 1.73233 +trainer/policy/normal/log_std Min -1.42532 +eval/num steps total 855446 +eval/num paths total 858 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117393 +eval/Actions Std 0.908843 +eval/Actions Max 0.999981 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52623 +time/logging (s) 0.00382592 +time/sampling batch (s) 0.27354 +time/saving (s) 0.00343354 +time/training (s) 7.04095 +time/epoch (s) 9.84798 +time/total (s) 8890.26 +Epoch -143 +---------------------------------- --------------- +2022-05-10 15:39:10.417039 PDT | [1] Epoch -142 finished +---------------------------------- --------------- +epoch -142 +replay_buffer/size 999033 +trainer/num train calls 859000 +trainer/Policy Loss -19.4701 +trainer/Log Pis Mean 25.2353 +trainer/Log Pis Std 13.538 +trainer/Log Pis Max 67.784 +trainer/Log Pis Min -12.9205 +trainer/policy/mean Mean -0.0543229 +trainer/policy/mean Std 0.910177 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83907 +trainer/policy/normal/std Std 0.681226 +trainer/policy/normal/std Max 6.38146 +trainer/policy/normal/std Min 0.215282 +trainer/policy/normal/log_std Mean 1.00153 +trainer/policy/normal/log_std Std 0.329661 +trainer/policy/normal/log_std Max 1.8534 +trainer/policy/normal/log_std Min -1.53581 +eval/num steps total 856446 +eval/num paths total 859 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.144835 +eval/Actions Std 0.892675 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53346 +time/logging (s) 0.00423125 +time/sampling batch (s) 0.52361 +time/saving (s) 0.00400294 +time/training (s) 7.14395 +time/epoch (s) 10.2093 +time/total (s) 8900.47 +Epoch -142 +---------------------------------- --------------- +2022-05-10 15:39:20.346078 PDT | [1] Epoch -141 finished +---------------------------------- --------------- +epoch -141 +replay_buffer/size 999033 +trainer/num train calls 860000 +trainer/Policy Loss -20.9726 +trainer/Log Pis Mean 24.3921 +trainer/Log Pis Std 12.9979 +trainer/Log Pis Max 70.4991 +trainer/Log Pis Min -4.86224 +trainer/policy/mean Mean -0.0284591 +trainer/policy/mean Std 0.907196 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82983 +trainer/policy/normal/std Std 0.682636 +trainer/policy/normal/std Max 6.24225 +trainer/policy/normal/std Min 0.243567 +trainer/policy/normal/log_std Mean 0.996236 +trainer/policy/normal/log_std Std 0.340345 +trainer/policy/normal/log_std Max 1.83134 +trainer/policy/normal/log_std Min -1.41236 +eval/num steps total 857446 +eval/num paths total 860 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.156734 +eval/Actions Std 0.918045 +eval/Actions Max 0.999983 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30785 +time/logging (s) 0.00381469 +time/sampling batch (s) 0.273979 +time/saving (s) 0.0035561 +time/training (s) 7.31442 +time/epoch (s) 9.90362 +time/total (s) 8910.38 +Epoch -141 +---------------------------------- --------------- +2022-05-10 15:39:31.354665 PDT | [1] Epoch -140 finished +---------------------------------- --------------- +epoch -140 +replay_buffer/size 999033 +trainer/num train calls 861000 +trainer/Policy Loss -19.2645 +trainer/Log Pis Mean 25.6045 +trainer/Log Pis Std 14.048 +trainer/Log Pis Max 70.3833 +trainer/Log Pis Min -10.9276 +trainer/policy/mean Mean -0.0117908 +trainer/policy/mean Std 0.908664 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.75537 +trainer/policy/normal/std Std 0.680893 +trainer/policy/normal/std Max 6.29485 +trainer/policy/normal/std Min 0.261193 +trainer/policy/normal/log_std Mean 0.969054 +trainer/policy/normal/log_std Std 0.339008 +trainer/policy/normal/log_std Max 1.83973 +trainer/policy/normal/log_std Min -1.3425 +eval/num steps total 858446 +eval/num paths total 861 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.066633 +eval/Actions Std 0.919965 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40912 +time/logging (s) 0.00370931 +time/sampling batch (s) 0.274173 +time/saving (s) 0.00339728 +time/training (s) 8.29371 +time/epoch (s) 10.9841 +time/total (s) 8921.37 +Epoch -140 +---------------------------------- --------------- +2022-05-10 15:39:42.705412 PDT | [1] Epoch -139 finished +---------------------------------- --------------- +epoch -139 +replay_buffer/size 999033 +trainer/num train calls 862000 +trainer/Policy Loss -20.634 +trainer/Log Pis Mean 24.5745 +trainer/Log Pis Std 12.7772 +trainer/Log Pis Max 65.1372 +trainer/Log Pis Min -3.64259 +trainer/policy/mean Mean -0.0299277 +trainer/policy/mean Std 0.910879 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.74496 +trainer/policy/normal/std Std 0.646207 +trainer/policy/normal/std Max 6.36574 +trainer/policy/normal/std Min 0.299692 +trainer/policy/normal/log_std Mean 0.96994 +trainer/policy/normal/log_std Std 0.318581 +trainer/policy/normal/log_std Max 1.85093 +trainer/policy/normal/log_std Min -1.205 +eval/num steps total 859446 +eval/num paths total 862 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0288779 +eval/Actions Std 0.909883 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76037 +time/logging (s) 0.0038069 +time/sampling batch (s) 0.52505 +time/saving (s) 0.00348714 +time/training (s) 8.03354 +time/epoch (s) 11.3263 +time/total (s) 8932.7 +Epoch -139 +---------------------------------- --------------- +2022-05-10 15:39:52.578821 PDT | [1] Epoch -138 finished +---------------------------------- --------------- +epoch -138 +replay_buffer/size 999033 +trainer/num train calls 863000 +trainer/Policy Loss -20.3213 +trainer/Log Pis Mean 23.4951 +trainer/Log Pis Std 13.4606 +trainer/Log Pis Max 74.4328 +trainer/Log Pis Min -8.17841 +trainer/policy/mean Mean -0.033731 +trainer/policy/mean Std 0.90525 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.76755 +trainer/policy/normal/std Std 0.651123 +trainer/policy/normal/std Max 6.11855 +trainer/policy/normal/std Min 0.296393 +trainer/policy/normal/log_std Mean 0.978384 +trainer/policy/normal/log_std Std 0.317456 +trainer/policy/normal/log_std Max 1.81133 +trainer/policy/normal/log_std Min -1.21607 +eval/num steps total 860446 +eval/num paths total 863 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0275529 +eval/Actions Std 0.920732 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45872 +time/logging (s) 0.00376081 +time/sampling batch (s) 0.775593 +time/saving (s) 0.0034731 +time/training (s) 6.60728 +time/epoch (s) 9.84882 +time/total (s) 8942.55 +Epoch -138 +---------------------------------- --------------- +2022-05-10 15:40:02.965327 PDT | [1] Epoch -137 finished +---------------------------------- --------------- +epoch -137 +replay_buffer/size 999033 +trainer/num train calls 864000 +trainer/Policy Loss -19.5478 +trainer/Log Pis Mean 23.4613 +trainer/Log Pis Std 13.7248 +trainer/Log Pis Max 70.5961 +trainer/Log Pis Min -6.85869 +trainer/policy/mean Mean -0.0222271 +trainer/policy/mean Std 0.901767 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79681 +trainer/policy/normal/std Std 0.678058 +trainer/policy/normal/std Max 6.9297 +trainer/policy/normal/std Min 0.242565 +trainer/policy/normal/log_std Mean 0.986489 +trainer/policy/normal/log_std Std 0.327659 +trainer/policy/normal/log_std Max 1.93582 +trainer/policy/normal/log_std Min -1.41649 +eval/num steps total 861446 +eval/num paths total 864 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0126214 +eval/Actions Std 0.902116 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57667 +time/logging (s) 0.00428432 +time/sampling batch (s) 0.774243 +time/saving (s) 0.00417295 +time/training (s) 7.00305 +time/epoch (s) 10.3624 +time/total (s) 8952.92 +Epoch -137 +---------------------------------- --------------- +2022-05-10 15:40:13.213367 PDT | [1] Epoch -136 finished +---------------------------------- --------------- +epoch -136 +replay_buffer/size 999033 +trainer/num train calls 865000 +trainer/Policy Loss -19.2805 +trainer/Log Pis Mean 23.9319 +trainer/Log Pis Std 12.9007 +trainer/Log Pis Max 66.8243 +trainer/Log Pis Min -8.25023 +trainer/policy/mean Mean -0.0387911 +trainer/policy/mean Std 0.905533 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77134 +trainer/policy/normal/std Std 0.686317 +trainer/policy/normal/std Max 5.83333 +trainer/policy/normal/std Min 0.27077 +trainer/policy/normal/log_std Mean 0.974938 +trainer/policy/normal/log_std Std 0.337274 +trainer/policy/normal/log_std Max 1.76359 +trainer/policy/normal/log_std Min -1.30649 +eval/num steps total 862446 +eval/num paths total 865 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200976 +eval/Actions Std 0.926057 +eval/Actions Max 0.999985 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58156 +time/logging (s) 0.00368743 +time/sampling batch (s) 0.774705 +time/saving (s) 0.00348818 +time/training (s) 6.85898 +time/epoch (s) 10.2224 +time/total (s) 8963.14 +Epoch -136 +---------------------------------- --------------- +2022-05-10 15:40:22.850625 PDT | [1] Epoch -135 finished +---------------------------------- --------------- +epoch -135 +replay_buffer/size 999033 +trainer/num train calls 866000 +trainer/Policy Loss -20.5774 +trainer/Log Pis Mean 24.2108 +trainer/Log Pis Std 13.9107 +trainer/Log Pis Max 72.7659 +trainer/Log Pis Min -6.81532 +trainer/policy/mean Mean -0.065989 +trainer/policy/mean Std 0.906142 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82874 +trainer/policy/normal/std Std 0.667005 +trainer/policy/normal/std Max 6.2421 +trainer/policy/normal/std Min 0.303381 +trainer/policy/normal/log_std Mean 1.00105 +trainer/policy/normal/log_std Std 0.31184 +trainer/policy/normal/log_std Max 1.83132 +trainer/policy/normal/log_std Min -1.19277 +eval/num steps total 863446 +eval/num paths total 866 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.201766 +eval/Actions Std 0.824765 +eval/Actions Max 0.999992 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47758 +time/logging (s) 0.00378697 +time/sampling batch (s) 0.275366 +time/saving (s) 0.00337687 +time/training (s) 6.85278 +time/epoch (s) 9.61289 +time/total (s) 8972.76 +Epoch -135 +---------------------------------- --------------- +2022-05-10 15:40:34.013655 PDT | [1] Epoch -134 finished +---------------------------------- --------------- +epoch -134 +replay_buffer/size 999033 +trainer/num train calls 867000 +trainer/Policy Loss -19.2863 +trainer/Log Pis Mean 24.1088 +trainer/Log Pis Std 13.1146 +trainer/Log Pis Max 75.1369 +trainer/Log Pis Min -4.62833 +trainer/policy/mean Mean -0.0485793 +trainer/policy/mean Std 0.904452 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79614 +trainer/policy/normal/std Std 0.669483 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.285787 +trainer/policy/normal/log_std Mean 0.986446 +trainer/policy/normal/log_std Std 0.328533 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.25251 +eval/num steps total 864446 +eval/num paths total 867 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249205 +eval/Actions Std 0.85396 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50641 +time/logging (s) 0.00383002 +time/sampling batch (s) 0.27545 +time/saving (s) 0.00340787 +time/training (s) 8.34954 +time/epoch (s) 11.1386 +time/total (s) 8983.9 +Epoch -134 +---------------------------------- --------------- +2022-05-10 15:40:43.668932 PDT | [1] Epoch -133 finished +---------------------------------- --------------- +epoch -133 +replay_buffer/size 999033 +trainer/num train calls 868000 +trainer/Policy Loss -20.7467 +trainer/Log Pis Mean 24.6711 +trainer/Log Pis Std 13.8069 +trainer/Log Pis Max 68.2701 +trainer/Log Pis Min -6.87931 +trainer/policy/mean Mean -0.0424032 +trainer/policy/mean Std 0.904796 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.79364 +trainer/policy/normal/std Std 0.673439 +trainer/policy/normal/std Max 5.75313 +trainer/policy/normal/std Min 0.285486 +trainer/policy/normal/log_std Mean 0.984018 +trainer/policy/normal/log_std Std 0.336 +trainer/policy/normal/log_std Max 1.74974 +trainer/policy/normal/log_std Min -1.25356 +eval/num steps total 865446 +eval/num paths total 868 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0828048 +eval/Actions Std 0.920194 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61768 +time/logging (s) 0.00396474 +time/sampling batch (s) 0.274433 +time/saving (s) 0.00390999 +time/training (s) 6.73085 +time/epoch (s) 9.63084 +time/total (s) 8993.53 +Epoch -133 +---------------------------------- --------------- +2022-05-10 15:40:53.002535 PDT | [1] Epoch -132 finished +---------------------------------- --------------- +epoch -132 +replay_buffer/size 999033 +trainer/num train calls 869000 +trainer/Policy Loss -20.0829 +trainer/Log Pis Mean 25.5122 +trainer/Log Pis Std 13.7589 +trainer/Log Pis Max 67.176 +trainer/Log Pis Min -6.53834 +trainer/policy/mean Mean -0.0511685 +trainer/policy/mean Std 0.905335 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.7669 +trainer/policy/normal/std Std 0.677165 +trainer/policy/normal/std Max 6.15379 +trainer/policy/normal/std Min 0.201168 +trainer/policy/normal/log_std Mean 0.973492 +trainer/policy/normal/log_std Std 0.340297 +trainer/policy/normal/log_std Max 1.81707 +trainer/policy/normal/log_std Min -1.60361 +eval/num steps total 866446 +eval/num paths total 869 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0329896 +eval/Actions Std 0.908156 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69583 +time/logging (s) 0.00396604 +time/sampling batch (s) 0.273357 +time/saving (s) 0.00339725 +time/training (s) 6.33235 +time/epoch (s) 9.3089 +time/total (s) 9002.84 +Epoch -132 +---------------------------------- --------------- +2022-05-10 15:41:03.817850 PDT | [1] Epoch -131 finished +---------------------------------- --------------- +epoch -131 +replay_buffer/size 999033 +trainer/num train calls 870000 +trainer/Policy Loss -19.3248 +trainer/Log Pis Mean 25.0697 +trainer/Log Pis Std 13.4264 +trainer/Log Pis Max 68.256 +trainer/Log Pis Min -5.83087 +trainer/policy/mean Mean -0.036759 +trainer/policy/mean Std 0.911544 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8545 +trainer/policy/normal/std Std 0.684899 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.303948 +trainer/policy/normal/log_std Mean 1.00884 +trainer/policy/normal/log_std Std 0.317845 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.1909 +eval/num steps total 867446 +eval/num paths total 870 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0471653 +eval/Actions Std 0.912125 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74796 +time/logging (s) 0.00385086 +time/sampling batch (s) 0.274512 +time/saving (s) 0.00355375 +time/training (s) 7.76078 +time/epoch (s) 10.7907 +time/total (s) 9013.64 +Epoch -131 +---------------------------------- --------------- +2022-05-10 15:41:14.062219 PDT | [1] Epoch -130 finished +---------------------------------- --------------- +epoch -130 +replay_buffer/size 999033 +trainer/num train calls 871000 +trainer/Policy Loss -19.3178 +trainer/Log Pis Mean 25.3834 +trainer/Log Pis Std 12.6858 +trainer/Log Pis Max 60.072 +trainer/Log Pis Min -5.46744 +trainer/policy/mean Mean -0.0418521 +trainer/policy/mean Std 0.908529 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81747 +trainer/policy/normal/std Std 0.671637 +trainer/policy/normal/std Max 6.46883 +trainer/policy/normal/std Min 0.30478 +trainer/policy/normal/log_std Mean 0.99451 +trainer/policy/normal/log_std Std 0.32581 +trainer/policy/normal/log_std Max 1.86699 +trainer/policy/normal/log_std Min -1.18816 +eval/num steps total 868446 +eval/num paths total 871 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.51555 +eval/Actions Std 0.768252 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71078 +time/logging (s) 0.00402217 +time/sampling batch (s) 0.280951 +time/saving (s) 0.00353079 +time/training (s) 7.22029 +time/epoch (s) 10.2196 +time/total (s) 9023.86 +Epoch -130 +---------------------------------- --------------- +2022-05-10 15:41:23.686160 PDT | [1] Epoch -129 finished +---------------------------------- --------------- +epoch -129 +replay_buffer/size 999033 +trainer/num train calls 872000 +trainer/Policy Loss -19.4669 +trainer/Log Pis Mean 24.1766 +trainer/Log Pis Std 12.9117 +trainer/Log Pis Max 79.4046 +trainer/Log Pis Min -4.33784 +trainer/policy/mean Mean -0.0437838 +trainer/policy/mean Std 0.904403 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76142 +trainer/policy/normal/std Std 0.677874 +trainer/policy/normal/std Max 6.47807 +trainer/policy/normal/std Min 0.236661 +trainer/policy/normal/log_std Mean 0.971736 +trainer/policy/normal/log_std Std 0.337704 +trainer/policy/normal/log_std Max 1.86842 +trainer/policy/normal/log_std Min -1.44113 +eval/num steps total 869446 +eval/num paths total 872 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206539 +eval/Actions Std 0.905769 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57077 +time/logging (s) 0.00378283 +time/sampling batch (s) 0.28125 +time/saving (s) 0.00347493 +time/training (s) 6.73913 +time/epoch (s) 9.59841 +time/total (s) 9033.46 +Epoch -129 +---------------------------------- --------------- +2022-05-10 15:41:32.768218 PDT | [1] Epoch -128 finished +---------------------------------- --------------- +epoch -128 +replay_buffer/size 999033 +trainer/num train calls 873000 +trainer/Policy Loss -18.8971 +trainer/Log Pis Mean 24.9884 +trainer/Log Pis Std 13.0401 +trainer/Log Pis Max 63.4041 +trainer/Log Pis Min -4.32079 +trainer/policy/mean Mean -0.0534238 +trainer/policy/mean Std 0.906453 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78827 +trainer/policy/normal/std Std 0.66812 +trainer/policy/normal/std Max 5.53948 +trainer/policy/normal/std Min 0.270105 +trainer/policy/normal/log_std Mean 0.984706 +trainer/policy/normal/log_std Std 0.32114 +trainer/policy/normal/log_std Max 1.7119 +trainer/policy/normal/log_std Min -1.30895 +eval/num steps total 870446 +eval/num paths total 873 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227942 +eval/Actions Std 0.755605 +eval/Actions Max 0.999987 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68903 +time/logging (s) 0.00416745 +time/sampling batch (s) 0.277542 +time/saving (s) 0.00379455 +time/training (s) 6.08311 +time/epoch (s) 9.05764 +time/total (s) 9042.52 +Epoch -128 +---------------------------------- --------------- +2022-05-10 15:41:43.464655 PDT | [1] Epoch -127 finished +---------------------------------- --------------- +epoch -127 +replay_buffer/size 999033 +trainer/num train calls 874000 +trainer/Policy Loss -19.6575 +trainer/Log Pis Mean 23.6296 +trainer/Log Pis Std 12.8234 +trainer/Log Pis Max 71.2284 +trainer/Log Pis Min -11.2757 +trainer/policy/mean Mean -0.0295844 +trainer/policy/mean Std 0.90862 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8205 +trainer/policy/normal/std Std 0.652109 +trainer/policy/normal/std Max 7.04669 +trainer/policy/normal/std Min 0.285106 +trainer/policy/normal/log_std Mean 0.999869 +trainer/policy/normal/log_std Std 0.304294 +trainer/policy/normal/log_std Max 1.95256 +trainer/policy/normal/log_std Min -1.2549 +eval/num steps total 871446 +eval/num paths total 874 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.202185 +eval/Actions Std 0.877985 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47317 +time/logging (s) 0.00388234 +time/sampling batch (s) 0.27903 +time/saving (s) 0.00373251 +time/training (s) 7.91125 +time/epoch (s) 10.6711 +time/total (s) 9053.2 +Epoch -127 +---------------------------------- --------------- +2022-05-10 15:41:53.849528 PDT | [1] Epoch -126 finished +---------------------------------- --------------- +epoch -126 +replay_buffer/size 999033 +trainer/num train calls 875000 +trainer/Policy Loss -21.2144 +trainer/Log Pis Mean 25.5292 +trainer/Log Pis Std 13.469 +trainer/Log Pis Max 70.9772 +trainer/Log Pis Min -8.03493 +trainer/policy/mean Mean -0.029845 +trainer/policy/mean Std 0.910155 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8353 +trainer/policy/normal/std Std 0.67427 +trainer/policy/normal/std Max 5.59145 +trainer/policy/normal/std Min 0.341533 +trainer/policy/normal/log_std Mean 1.00004 +trainer/policy/normal/log_std Std 0.330492 +trainer/policy/normal/log_std Max 1.72124 +trainer/policy/normal/log_std Min -1.07431 +eval/num steps total 872446 +eval/num paths total 875 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.338132 +eval/Actions Std 0.804193 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52682 +time/logging (s) 0.00372466 +time/sampling batch (s) 0.27903 +time/saving (s) 0.00340529 +time/training (s) 7.54667 +time/epoch (s) 10.3596 +time/total (s) 9063.56 +Epoch -126 +---------------------------------- --------------- +2022-05-10 15:42:04.547101 PDT | [1] Epoch -125 finished +---------------------------------- --------------- +epoch -125 +replay_buffer/size 999033 +trainer/num train calls 876000 +trainer/Policy Loss -20.6558 +trainer/Log Pis Mean 24.6653 +trainer/Log Pis Std 13.9716 +trainer/Log Pis Max 67.8295 +trainer/Log Pis Min -7.70362 +trainer/policy/mean Mean -0.052799 +trainer/policy/mean Std 0.908791 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7786 +trainer/policy/normal/std Std 0.664944 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.295029 +trainer/policy/normal/log_std Mean 0.980939 +trainer/policy/normal/log_std Std 0.32306 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22068 +eval/num steps total 873446 +eval/num paths total 876 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115068 +eval/Actions Std 0.913033 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50537 +time/logging (s) 0.00377296 +time/sampling batch (s) 0.525137 +time/saving (s) 0.00346396 +time/training (s) 7.63538 +time/epoch (s) 10.6731 +time/total (s) 9074.24 +Epoch -125 +---------------------------------- --------------- +2022-05-10 15:42:13.493145 PDT | [1] Epoch -124 finished +---------------------------------- --------------- +epoch -124 +replay_buffer/size 999033 +trainer/num train calls 877000 +trainer/Policy Loss -19.2007 +trainer/Log Pis Mean 24.3345 +trainer/Log Pis Std 13.1482 +trainer/Log Pis Max 65.8636 +trainer/Log Pis Min -5.80391 +trainer/policy/mean Mean -0.0550525 +trainer/policy/mean Std 0.901495 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.76332 +trainer/policy/normal/std Std 0.648593 +trainer/policy/normal/std Max 4.91871 +trainer/policy/normal/std Min 0.254528 +trainer/policy/normal/log_std Mean 0.975497 +trainer/policy/normal/log_std Std 0.325616 +trainer/policy/normal/log_std Max 1.59305 +trainer/policy/normal/log_std Min -1.36834 +eval/num steps total 874446 +eval/num paths total 877 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0454747 +eval/Actions Std 0.900921 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73401 +time/logging (s) 0.00373072 +time/sampling batch (s) 0.273625 +time/saving (s) 0.00346873 +time/training (s) 5.90681 +time/epoch (s) 8.92165 +time/total (s) 9083.16 +Epoch -124 +---------------------------------- --------------- +2022-05-10 15:42:23.555822 PDT | [1] Epoch -123 finished +---------------------------------- --------------- +epoch -123 +replay_buffer/size 999033 +trainer/num train calls 878000 +trainer/Policy Loss -20.0569 +trainer/Log Pis Mean 24.2058 +trainer/Log Pis Std 13.4188 +trainer/Log Pis Max 79.188 +trainer/Log Pis Min -9.8543 +trainer/policy/mean Mean -0.0409617 +trainer/policy/mean Std 0.90939 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79286 +trainer/policy/normal/std Std 0.676984 +trainer/policy/normal/std Max 5.93922 +trainer/policy/normal/std Min 0.232041 +trainer/policy/normal/log_std Mean 0.983211 +trainer/policy/normal/log_std Std 0.338792 +trainer/policy/normal/log_std Max 1.78158 +trainer/policy/normal/log_std Min -1.46084 +eval/num steps total 875446 +eval/num paths total 878 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0344887 +eval/Actions Std 0.904588 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65917 +time/logging (s) 0.00418399 +time/sampling batch (s) 0.524832 +time/saving (s) 0.00406298 +time/training (s) 6.84641 +time/epoch (s) 10.0387 +time/total (s) 9093.2 +Epoch -123 +---------------------------------- --------------- +2022-05-10 15:42:34.277673 PDT | [1] Epoch -122 finished +---------------------------------- --------------- +epoch -122 +replay_buffer/size 999033 +trainer/num train calls 879000 +trainer/Policy Loss -19.3841 +trainer/Log Pis Mean 24.2417 +trainer/Log Pis Std 13.3382 +trainer/Log Pis Max 79.2259 +trainer/Log Pis Min -9.90411 +trainer/policy/mean Mean -0.066087 +trainer/policy/mean Std 0.901731 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77394 +trainer/policy/normal/std Std 0.689404 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.236069 +trainer/policy/normal/log_std Mean 0.975037 +trainer/policy/normal/log_std Std 0.343132 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.44363 +eval/num steps total 876446 +eval/num paths total 879 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216971 +eval/Actions Std 0.819145 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66503 +time/logging (s) 0.00378047 +time/sampling batch (s) 0.525693 +time/saving (s) 0.0034215 +time/training (s) 7.49866 +time/epoch (s) 10.6966 +time/total (s) 9103.9 +Epoch -122 +---------------------------------- --------------- +2022-05-10 15:42:44.973523 PDT | [1] Epoch -121 finished +---------------------------------- --------------- +epoch -121 +replay_buffer/size 999033 +trainer/num train calls 880000 +trainer/Policy Loss -19.2245 +trainer/Log Pis Mean 24.077 +trainer/Log Pis Std 13.6249 +trainer/Log Pis Max 78.2899 +trainer/Log Pis Min -10.9883 +trainer/policy/mean Mean -0.0345202 +trainer/policy/mean Std 0.906234 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75854 +trainer/policy/normal/std Std 0.657109 +trainer/policy/normal/std Max 5.87032 +trainer/policy/normal/std Min 0.298047 +trainer/policy/normal/log_std Mean 0.973617 +trainer/policy/normal/log_std Std 0.323514 +trainer/policy/normal/log_std Max 1.76991 +trainer/policy/normal/log_std Min -1.2105 +eval/num steps total 877446 +eval/num paths total 880 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.352193 +eval/Actions Std 0.895068 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6792 +time/logging (s) 0.00380642 +time/sampling batch (s) 0.286409 +time/saving (s) 0.00365767 +time/training (s) 7.69789 +time/epoch (s) 10.671 +time/total (s) 9114.58 +Epoch -121 +---------------------------------- --------------- +2022-05-10 15:42:56.289055 PDT | [1] Epoch -120 finished +---------------------------------- --------------- +epoch -120 +replay_buffer/size 999033 +trainer/num train calls 881000 +trainer/Policy Loss -20.0973 +trainer/Log Pis Mean 24.8882 +trainer/Log Pis Std 12.7631 +trainer/Log Pis Max 63.104 +trainer/Log Pis Min -5.64319 +trainer/policy/mean Mean -0.029159 +trainer/policy/mean Std 0.909103 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.73455 +trainer/policy/normal/std Std 0.666595 +trainer/policy/normal/std Max 6.41139 +trainer/policy/normal/std Min 0.305046 +trainer/policy/normal/log_std Mean 0.962377 +trainer/policy/normal/log_std Std 0.334835 +trainer/policy/normal/log_std Max 1.85808 +trainer/policy/normal/log_std Min -1.18729 +eval/num steps total 878446 +eval/num paths total 881 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00160117 +eval/Actions Std 0.90536 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87657 +time/logging (s) 0.00378007 +time/sampling batch (s) 0.580335 +time/saving (s) 0.00349855 +time/training (s) 7.82543 +time/epoch (s) 11.2896 +time/total (s) 9125.87 +Epoch -120 +---------------------------------- --------------- +2022-05-10 15:43:05.915440 PDT | [1] Epoch -119 finished +---------------------------------- --------------- +epoch -119 +replay_buffer/size 999033 +trainer/num train calls 882000 +trainer/Policy Loss -19.9021 +trainer/Log Pis Mean 24.0063 +trainer/Log Pis Std 12.9939 +trainer/Log Pis Max 63.8669 +trainer/Log Pis Min -8.76358 +trainer/policy/mean Mean -0.0407191 +trainer/policy/mean Std 0.908388 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81976 +trainer/policy/normal/std Std 0.672713 +trainer/policy/normal/std Max 6.12567 +trainer/policy/normal/std Min 0.255692 +trainer/policy/normal/log_std Mean 0.993515 +trainer/policy/normal/log_std Std 0.33764 +trainer/policy/normal/log_std Max 1.81249 +trainer/policy/normal/log_std Min -1.36378 +eval/num steps total 879446 +eval/num paths total 882 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00548224 +eval/Actions Std 0.910866 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51328 +time/logging (s) 0.00383217 +time/sampling batch (s) 0.27846 +time/saving (s) 0.00350058 +time/training (s) 6.80258 +time/epoch (s) 9.60166 +time/total (s) 9135.47 +Epoch -119 +---------------------------------- --------------- +2022-05-10 15:43:15.321631 PDT | [1] Epoch -118 finished +---------------------------------- --------------- +epoch -118 +replay_buffer/size 999033 +trainer/num train calls 883000 +trainer/Policy Loss -20.1566 +trainer/Log Pis Mean 24.4323 +trainer/Log Pis Std 12.6746 +trainer/Log Pis Max 65.5936 +trainer/Log Pis Min -4.4324 +trainer/policy/mean Mean -0.0246621 +trainer/policy/mean Std 0.910068 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76548 +trainer/policy/normal/std Std 0.682897 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.228358 +trainer/policy/normal/log_std Mean 0.972626 +trainer/policy/normal/log_std Std 0.341648 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.47684 +eval/num steps total 880446 +eval/num paths total 883 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.197788 +eval/Actions Std 0.889209 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65651 +time/logging (s) 0.00380465 +time/sampling batch (s) 0.279093 +time/saving (s) 0.00345735 +time/training (s) 6.43837 +time/epoch (s) 9.38124 +time/total (s) 9144.86 +Epoch -118 +---------------------------------- --------------- +2022-05-10 15:43:26.066168 PDT | [1] Epoch -117 finished +---------------------------------- --------------- +epoch -117 +replay_buffer/size 999033 +trainer/num train calls 884000 +trainer/Policy Loss -21.2607 +trainer/Log Pis Mean 25.2019 +trainer/Log Pis Std 13.2881 +trainer/Log Pis Max 72.2982 +trainer/Log Pis Min -5.8258 +trainer/policy/mean Mean -0.051224 +trainer/policy/mean Std 0.905234 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.829 +trainer/policy/normal/std Std 0.689398 +trainer/policy/normal/std Max 6.28948 +trainer/policy/normal/std Min 0.277102 +trainer/policy/normal/log_std Mean 0.995685 +trainer/policy/normal/log_std Std 0.339919 +trainer/policy/normal/log_std Max 1.83888 +trainer/policy/normal/log_std Min -1.28337 +eval/num steps total 881446 +eval/num paths total 884 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0886262 +eval/Actions Std 0.903571 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.19723 +time/logging (s) 0.00486066 +time/sampling batch (s) 0.289884 +time/saving (s) 0.00459266 +time/training (s) 8.22358 +time/epoch (s) 10.7201 +time/total (s) 9155.58 +Epoch -117 +---------------------------------- --------------- +2022-05-10 15:43:36.653263 PDT | [1] Epoch -116 finished +---------------------------------- --------------- +epoch -116 +replay_buffer/size 999033 +trainer/num train calls 885000 +trainer/Policy Loss -20.4166 +trainer/Log Pis Mean 23.8431 +trainer/Log Pis Std 13.1977 +trainer/Log Pis Max 65.8795 +trainer/Log Pis Min -10.223 +trainer/policy/mean Mean -0.0501611 +trainer/policy/mean Std 0.908692 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81533 +trainer/policy/normal/std Std 0.691198 +trainer/policy/normal/std Max 5.97881 +trainer/policy/normal/std Min 0.228112 +trainer/policy/normal/log_std Mean 0.989921 +trainer/policy/normal/log_std Std 0.344167 +trainer/policy/normal/log_std Max 1.78822 +trainer/policy/normal/log_std Min -1.47792 +eval/num steps total 882446 +eval/num paths total 885 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.622887 +eval/Actions Std 0.752129 +eval/Actions Max 0.999975 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58088 +time/logging (s) 0.003717 +time/sampling batch (s) 0.776304 +time/saving (s) 0.00360829 +time/training (s) 7.1961 +time/epoch (s) 10.5606 +time/total (s) 9166.14 +Epoch -116 +---------------------------------- --------------- +2022-05-10 15:43:47.659774 PDT | [1] Epoch -115 finished +---------------------------------- --------------- +epoch -115 +replay_buffer/size 999033 +trainer/num train calls 886000 +trainer/Policy Loss -20.2767 +trainer/Log Pis Mean 24.8835 +trainer/Log Pis Std 13.6649 +trainer/Log Pis Max 65.328 +trainer/Log Pis Min -8.66026 +trainer/policy/mean Mean -0.049595 +trainer/policy/mean Std 0.907567 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80966 +trainer/policy/normal/std Std 0.675199 +trainer/policy/normal/std Max 5.81557 +trainer/policy/normal/std Min 0.293224 +trainer/policy/normal/log_std Mean 0.990049 +trainer/policy/normal/log_std Std 0.334468 +trainer/policy/normal/log_std Max 1.76054 +trainer/policy/normal/log_std Min -1.22682 +eval/num steps total 883446 +eval/num paths total 886 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0156917 +eval/Actions Std 0.908406 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64736 +time/logging (s) 0.00376682 +time/sampling batch (s) 0.781214 +time/saving (s) 0.00349297 +time/training (s) 7.54566 +time/epoch (s) 10.9815 +time/total (s) 9177.13 +Epoch -115 +---------------------------------- --------------- +2022-05-10 15:43:57.935959 PDT | [1] Epoch -114 finished +---------------------------------- --------------- +epoch -114 +replay_buffer/size 999033 +trainer/num train calls 887000 +trainer/Policy Loss -18.2775 +trainer/Log Pis Mean 24.0775 +trainer/Log Pis Std 13.3481 +trainer/Log Pis Max 76.6699 +trainer/Log Pis Min -5.85679 +trainer/policy/mean Mean -0.0236076 +trainer/policy/mean Std 0.909518 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.80418 +trainer/policy/normal/std Std 0.691984 +trainer/policy/normal/std Max 6.76884 +trainer/policy/normal/std Min 0.250825 +trainer/policy/normal/log_std Mean 0.985881 +trainer/policy/normal/log_std Std 0.344332 +trainer/policy/normal/log_std Max 1.91233 +trainer/policy/normal/log_std Min -1.383 +eval/num steps total 884446 +eval/num paths total 887 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124442 +eval/Actions Std 0.947581 +eval/Actions Max 0.99999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51489 +time/logging (s) 0.00376427 +time/sampling batch (s) 0.531598 +time/saving (s) 0.00342593 +time/training (s) 7.19733 +time/epoch (s) 10.251 +time/total (s) 9187.38 +Epoch -114 +---------------------------------- --------------- +2022-05-10 15:44:07.615097 PDT | [1] Epoch -113 finished +---------------------------------- --------------- +epoch -113 +replay_buffer/size 999033 +trainer/num train calls 888000 +trainer/Policy Loss -18.8822 +trainer/Log Pis Mean 24.4559 +trainer/Log Pis Std 12.5884 +trainer/Log Pis Max 82.529 +trainer/Log Pis Min -4.46082 +trainer/policy/mean Mean -0.0413583 +trainer/policy/mean Std 0.908617 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79408 +trainer/policy/normal/std Std 0.646199 +trainer/policy/normal/std Max 5.69371 +trainer/policy/normal/std Min 0.294437 +trainer/policy/normal/log_std Mean 0.989514 +trainer/policy/normal/log_std Std 0.309491 +trainer/policy/normal/log_std Max 1.73936 +trainer/policy/normal/log_std Min -1.22269 +eval/num steps total 885446 +eval/num paths total 888 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132062 +eval/Actions Std 0.915411 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51517 +time/logging (s) 0.00376051 +time/sampling batch (s) 0.283505 +time/saving (s) 0.00338871 +time/training (s) 6.84809 +time/epoch (s) 9.65391 +time/total (s) 9197.04 +Epoch -113 +---------------------------------- --------------- +2022-05-10 15:44:18.217964 PDT | [1] Epoch -112 finished +---------------------------------- --------------- +epoch -112 +replay_buffer/size 999033 +trainer/num train calls 889000 +trainer/Policy Loss -19.2717 +trainer/Log Pis Mean 25.0665 +trainer/Log Pis Std 13.2759 +trainer/Log Pis Max 72.1516 +trainer/Log Pis Min -5.01292 +trainer/policy/mean Mean -0.0153699 +trainer/policy/mean Std 0.903533 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77042 +trainer/policy/normal/std Std 0.699198 +trainer/policy/normal/std Max 5.59286 +trainer/policy/normal/std Min 0.196472 +trainer/policy/normal/log_std Mean 0.970282 +trainer/policy/normal/log_std Std 0.359053 +trainer/policy/normal/log_std Max 1.72149 +trainer/policy/normal/log_std Min -1.62724 +eval/num steps total 886446 +eval/num paths total 889 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0029406 +eval/Actions Std 0.894921 +eval/Actions Max 1 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67392 +time/logging (s) 0.00369742 +time/sampling batch (s) 0.529131 +time/saving (s) 0.0034231 +time/training (s) 7.36768 +time/epoch (s) 10.5779 +time/total (s) 9207.62 +Epoch -112 +---------------------------------- --------------- +2022-05-10 15:44:29.207710 PDT | [1] Epoch -111 finished +---------------------------------- --------------- +epoch -111 +replay_buffer/size 999033 +trainer/num train calls 890000 +trainer/Policy Loss -19.1008 +trainer/Log Pis Mean 24.3003 +trainer/Log Pis Std 13.0138 +trainer/Log Pis Max 61.7114 +trainer/Log Pis Min -11.5812 +trainer/policy/mean Mean -0.0133767 +trainer/policy/mean Std 0.903852 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82126 +trainer/policy/normal/std Std 0.687119 +trainer/policy/normal/std Max 6.62771 +trainer/policy/normal/std Min 0.208252 +trainer/policy/normal/log_std Mean 0.994031 +trainer/policy/normal/log_std Std 0.333228 +trainer/policy/normal/log_std Max 1.89126 +trainer/policy/normal/log_std Min -1.56901 +eval/num steps total 887446 +eval/num paths total 890 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122997 +eval/Actions Std 0.88642 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42325 +time/logging (s) 0.00413823 +time/sampling batch (s) 0.279869 +time/saving (s) 0.00387157 +time/training (s) 8.25393 +time/epoch (s) 10.9651 +time/total (s) 9218.59 +Epoch -111 +---------------------------------- --------------- +2022-05-10 15:44:39.390647 PDT | [1] Epoch -110 finished +---------------------------------- --------------- +epoch -110 +replay_buffer/size 999033 +trainer/num train calls 891000 +trainer/Policy Loss -19.5453 +trainer/Log Pis Mean 24.976 +trainer/Log Pis Std 12.5377 +trainer/Log Pis Max 69.055 +trainer/Log Pis Min -4.35774 +trainer/policy/mean Mean -0.0318687 +trainer/policy/mean Std 0.905167 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81434 +trainer/policy/normal/std Std 0.676843 +trainer/policy/normal/std Max 6.4202 +trainer/policy/normal/std Min 0.295202 +trainer/policy/normal/log_std Mean 0.993561 +trainer/policy/normal/log_std Std 0.322705 +trainer/policy/normal/log_std Max 1.85945 +trainer/policy/normal/log_std Min -1.2201 +eval/num steps total 888446 +eval/num paths total 891 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.171745 +eval/Actions Std 0.919665 +eval/Actions Max 0.999989 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61567 +time/logging (s) 0.00380943 +time/sampling batch (s) 0.283 +time/saving (s) 0.00372882 +time/training (s) 7.25056 +time/epoch (s) 10.1568 +time/total (s) 9228.75 +Epoch -110 +---------------------------------- --------------- +2022-05-10 15:44:50.257033 PDT | [1] Epoch -109 finished +---------------------------------- --------------- +epoch -109 +replay_buffer/size 999033 +trainer/num train calls 892000 +trainer/Policy Loss -19.0732 +trainer/Log Pis Mean 25.3972 +trainer/Log Pis Std 13.3304 +trainer/Log Pis Max 65.5355 +trainer/Log Pis Min -5.73106 +trainer/policy/mean Mean -0.0531652 +trainer/policy/mean Std 0.908949 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79896 +trainer/policy/normal/std Std 0.65529 +trainer/policy/normal/std Max 6.39461 +trainer/policy/normal/std Min 0.284039 +trainer/policy/normal/log_std Mean 0.991463 +trainer/policy/normal/log_std Std 0.307347 +trainer/policy/normal/log_std Max 1.85546 +trainer/policy/normal/log_std Min -1.25864 +eval/num steps total 889446 +eval/num paths total 892 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.176459 +eval/Actions Std 0.887131 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65805 +time/logging (s) 0.00385321 +time/sampling batch (s) 0.530042 +time/saving (s) 0.00345752 +time/training (s) 7.6459 +time/epoch (s) 10.8413 +time/total (s) 9239.59 +Epoch -109 +---------------------------------- --------------- +2022-05-10 15:45:00.223488 PDT | [1] Epoch -108 finished +---------------------------------- --------------- +epoch -108 +replay_buffer/size 999033 +trainer/num train calls 893000 +trainer/Policy Loss -18.8595 +trainer/Log Pis Mean 24.6886 +trainer/Log Pis Std 12.6511 +trainer/Log Pis Max 60.8084 +trainer/Log Pis Min -10.9646 +trainer/policy/mean Mean -0.0350001 +trainer/policy/mean Std 0.904967 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.80872 +trainer/policy/normal/std Std 0.680577 +trainer/policy/normal/std Max 6.32607 +trainer/policy/normal/std Min 0.27427 +trainer/policy/normal/log_std Mean 0.988797 +trainer/policy/normal/log_std Std 0.339226 +trainer/policy/normal/log_std Max 1.84468 +trainer/policy/normal/log_std Min -1.29364 +eval/num steps total 890446 +eval/num paths total 893 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.248182 +eval/Actions Std 0.888137 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59178 +time/logging (s) 0.00383198 +time/sampling batch (s) 0.278917 +time/saving (s) 0.00348769 +time/training (s) 7.06329 +time/epoch (s) 9.94131 +time/total (s) 9249.54 +Epoch -108 +---------------------------------- --------------- +2022-05-10 15:45:10.308734 PDT | [1] Epoch -107 finished +---------------------------------- --------------- +epoch -107 +replay_buffer/size 999033 +trainer/num train calls 894000 +trainer/Policy Loss -19.1755 +trainer/Log Pis Mean 23.6633 +trainer/Log Pis Std 13.0182 +trainer/Log Pis Max 68.5379 +trainer/Log Pis Min -7.12791 +trainer/policy/mean Mean -0.0421301 +trainer/policy/mean Std 0.904182 +trainer/policy/mean Max 0.999971 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.73764 +trainer/policy/normal/std Std 0.682819 +trainer/policy/normal/std Max 7.18633 +trainer/policy/normal/std Min 0.252924 +trainer/policy/normal/log_std Mean 0.960057 +trainer/policy/normal/log_std Std 0.351944 +trainer/policy/normal/log_std Max 1.97218 +trainer/policy/normal/log_std Min -1.37467 +eval/num steps total 891446 +eval/num paths total 894 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0394681 +eval/Actions Std 0.920752 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54518 +time/logging (s) 0.00376597 +time/sampling batch (s) 0.27917 +time/saving (s) 0.00342686 +time/training (s) 7.22843 +time/epoch (s) 10.06 +time/total (s) 9259.6 +Epoch -107 +---------------------------------- --------------- +2022-05-10 15:45:20.507776 PDT | [1] Epoch -106 finished +---------------------------------- --------------- +epoch -106 +replay_buffer/size 999033 +trainer/num train calls 895000 +trainer/Policy Loss -19.2682 +trainer/Log Pis Mean 24.0146 +trainer/Log Pis Std 13.1093 +trainer/Log Pis Max 71.6746 +trainer/Log Pis Min -4.96066 +trainer/policy/mean Mean -0.00700944 +trainer/policy/mean Std 0.905758 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77322 +trainer/policy/normal/std Std 0.680191 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.245658 +trainer/policy/normal/log_std Mean 0.976294 +trainer/policy/normal/log_std Std 0.336684 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.40381 +eval/num steps total 892446 +eval/num paths total 895 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.320532 +eval/Actions Std 0.8583 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60347 +time/logging (s) 0.00384568 +time/sampling batch (s) 0.780182 +time/saving (s) 0.0034571 +time/training (s) 6.78308 +time/epoch (s) 10.174 +time/total (s) 9269.78 +Epoch -106 +---------------------------------- --------------- +2022-05-10 15:45:30.980125 PDT | [1] Epoch -105 finished +---------------------------------- --------------- +epoch -105 +replay_buffer/size 999033 +trainer/num train calls 896000 +trainer/Policy Loss -20.3207 +trainer/Log Pis Mean 24.5776 +trainer/Log Pis Std 13.6407 +trainer/Log Pis Max 77.4116 +trainer/Log Pis Min -7.63286 +trainer/policy/mean Mean -0.00645268 +trainer/policy/mean Std 0.907647 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83986 +trainer/policy/normal/std Std 0.704236 +trainer/policy/normal/std Max 6.19017 +trainer/policy/normal/std Min 0.240208 +trainer/policy/normal/log_std Mean 0.998405 +trainer/policy/normal/log_std Std 0.343082 +trainer/policy/normal/log_std Max 1.82296 +trainer/policy/normal/log_std Min -1.42625 +eval/num steps total 893446 +eval/num paths total 896 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.115069 +eval/Actions Std 0.916981 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57338 +time/logging (s) 0.00373933 +time/sampling batch (s) 0.283592 +time/saving (s) 0.00351684 +time/training (s) 7.58266 +time/epoch (s) 10.4469 +time/total (s) 9280.23 +Epoch -105 +---------------------------------- --------------- +2022-05-10 15:45:40.726764 PDT | [1] Epoch -104 finished +---------------------------------- --------------- +epoch -104 +replay_buffer/size 999033 +trainer/num train calls 897000 +trainer/Policy Loss -20.4684 +trainer/Log Pis Mean 24.8176 +trainer/Log Pis Std 13.2345 +trainer/Log Pis Max 69.3605 +trainer/Log Pis Min -6.32651 +trainer/policy/mean Mean -0.0521822 +trainer/policy/mean Std 0.908452 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79316 +trainer/policy/normal/std Std 0.680668 +trainer/policy/normal/std Max 5.6443 +trainer/policy/normal/std Min 0.239712 +trainer/policy/normal/log_std Mean 0.984069 +trainer/policy/normal/log_std Std 0.333339 +trainer/policy/normal/log_std Max 1.73065 +trainer/policy/normal/log_std Min -1.42832 +eval/num steps total 894446 +eval/num paths total 897 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0700273 +eval/Actions Std 0.914821 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.15631 +time/logging (s) 0.00398674 +time/sampling batch (s) 0.527648 +time/saving (s) 0.00384557 +time/training (s) 7.03012 +time/epoch (s) 9.7219 +time/total (s) 9289.95 +Epoch -104 +---------------------------------- --------------- +2022-05-10 15:45:49.647133 PDT | [1] Epoch -103 finished +---------------------------------- --------------- +epoch -103 +replay_buffer/size 999033 +trainer/num train calls 898000 +trainer/Policy Loss -21.0722 +trainer/Log Pis Mean 25.0814 +trainer/Log Pis Std 13.9686 +trainer/Log Pis Max 80.9848 +trainer/Log Pis Min -4.32794 +trainer/policy/mean Mean -0.0336938 +trainer/policy/mean Std 0.906589 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8849 +trainer/policy/normal/std Std 0.68346 +trainer/policy/normal/std Max 7.07966 +trainer/policy/normal/std Min 0.209415 +trainer/policy/normal/log_std Mean 1.01866 +trainer/policy/normal/log_std Std 0.325976 +trainer/policy/normal/log_std Max 1.95723 +trainer/policy/normal/log_std Min -1.56344 +eval/num steps total 895446 +eval/num paths total 898 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0290049 +eval/Actions Std 0.907187 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35132 +time/logging (s) 0.00389868 +time/sampling batch (s) 0.275857 +time/saving (s) 0.00360014 +time/training (s) 6.26054 +time/epoch (s) 8.89522 +time/total (s) 9298.85 +Epoch -103 +---------------------------------- --------------- +2022-05-10 15:46:00.128409 PDT | [1] Epoch -102 finished +---------------------------------- --------------- +epoch -102 +replay_buffer/size 999033 +trainer/num train calls 899000 +trainer/Policy Loss -19.7568 +trainer/Log Pis Mean 24.5992 +trainer/Log Pis Std 13.1956 +trainer/Log Pis Max 68.9862 +trainer/Log Pis Min -7.77848 +trainer/policy/mean Mean -0.0370329 +trainer/policy/mean Std 0.903227 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81276 +trainer/policy/normal/std Std 0.705318 +trainer/policy/normal/std Max 6.54238 +trainer/policy/normal/std Min 0.211533 +trainer/policy/normal/log_std Mean 0.985342 +trainer/policy/normal/log_std Std 0.36203 +trainer/policy/normal/log_std Max 1.8783 +trainer/policy/normal/log_std Min -1.55338 +eval/num steps total 896446 +eval/num paths total 899 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0564159 +eval/Actions Std 0.90134 +eval/Actions Max 0.999995 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70168 +time/logging (s) 0.00374698 +time/sampling batch (s) 0.529033 +time/saving (s) 0.00347964 +time/training (s) 7.21812 +time/epoch (s) 10.4561 +time/total (s) 9309.31 +Epoch -102 +---------------------------------- --------------- +2022-05-10 15:46:09.438873 PDT | [1] Epoch -101 finished +---------------------------------- --------------- +epoch -101 +replay_buffer/size 999033 +trainer/num train calls 900000 +trainer/Policy Loss -19.7107 +trainer/Log Pis Mean 23.3761 +trainer/Log Pis Std 13.2625 +trainer/Log Pis Max 62.444 +trainer/Log Pis Min -8.27983 +trainer/policy/mean Mean -0.0260639 +trainer/policy/mean Std 0.909368 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78185 +trainer/policy/normal/std Std 0.668199 +trainer/policy/normal/std Max 5.43217 +trainer/policy/normal/std Min 0.27787 +trainer/policy/normal/log_std Mean 0.980815 +trainer/policy/normal/log_std Std 0.329678 +trainer/policy/normal/log_std Max 1.69234 +trainer/policy/normal/log_std Min -1.2806 +eval/num steps total 897446 +eval/num paths total 900 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121227 +eval/Actions Std 0.880524 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3716 +time/logging (s) 0.00390748 +time/sampling batch (s) 0.275631 +time/saving (s) 0.00634991 +time/training (s) 6.62837 +time/epoch (s) 9.28586 +time/total (s) 9318.6 +Epoch -101 +---------------------------------- --------------- +2022-05-10 15:46:20.623519 PDT | [1] Epoch -100 finished +---------------------------------- --------------- +epoch -100 +replay_buffer/size 999033 +trainer/num train calls 901000 +trainer/Policy Loss -18.5782 +trainer/Log Pis Mean 24.6709 +trainer/Log Pis Std 12.8513 +trainer/Log Pis Max 61.3268 +trainer/Log Pis Min -7.71038 +trainer/policy/mean Mean -0.0312498 +trainer/policy/mean Std 0.907739 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.79499 +trainer/policy/normal/std Std 0.673882 +trainer/policy/normal/std Max 6.73656 +trainer/policy/normal/std Min 0.269709 +trainer/policy/normal/log_std Mean 0.98497 +trainer/policy/normal/log_std Std 0.333823 +trainer/policy/normal/log_std Max 1.90755 +trainer/policy/normal/log_std Min -1.31041 +eval/num steps total 898446 +eval/num paths total 901 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.330084 +eval/Actions Std 0.837173 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65172 +time/logging (s) 0.00384794 +time/sampling batch (s) 0.320065 +time/saving (s) 0.00383841 +time/training (s) 8.17944 +time/epoch (s) 11.1589 +time/total (s) 9329.76 +Epoch -100 +---------------------------------- --------------- +2022-05-10 15:46:31.930035 PDT | [1] Epoch -99 finished +---------------------------------- --------------- +epoch -99 +replay_buffer/size 999033 +trainer/num train calls 902000 +trainer/Policy Loss -20.0883 +trainer/Log Pis Mean 25.3033 +trainer/Log Pis Std 13.6553 +trainer/Log Pis Max 76.9465 +trainer/Log Pis Min -6.52808 +trainer/policy/mean Mean -0.0403256 +trainer/policy/mean Std 0.90621 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87284 +trainer/policy/normal/std Std 0.704532 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.242431 +trainer/policy/normal/log_std Mean 1.00975 +trainer/policy/normal/log_std Std 0.34759 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.41704 +eval/num steps total 899446 +eval/num paths total 902 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00458959 +eval/Actions Std 0.906273 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66633 +time/logging (s) 0.00377181 +time/sampling batch (s) 0.831951 +time/saving (s) 0.00364027 +time/training (s) 7.77488 +time/epoch (s) 11.2806 +time/total (s) 9341.04 +Epoch -99 +---------------------------------- --------------- +2022-05-10 15:46:42.821825 PDT | [1] Epoch -98 finished +---------------------------------- --------------- +epoch -98 +replay_buffer/size 999033 +trainer/num train calls 903000 +trainer/Policy Loss -18.6956 +trainer/Log Pis Mean 24.8042 +trainer/Log Pis Std 12.8522 +trainer/Log Pis Max 68.7756 +trainer/Log Pis Min -13.4045 +trainer/policy/mean Mean -0.041887 +trainer/policy/mean Std 0.90534 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82796 +trainer/policy/normal/std Std 0.682165 +trainer/policy/normal/std Max 6.46117 +trainer/policy/normal/std Min 0.253372 +trainer/policy/normal/log_std Mean 0.996342 +trainer/policy/normal/log_std Std 0.334803 +trainer/policy/normal/log_std Max 1.86581 +trainer/policy/normal/log_std Min -1.3729 +eval/num steps total 900446 +eval/num paths total 903 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105859 +eval/Actions Std 0.922148 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68617 +time/logging (s) 0.00493338 +time/sampling batch (s) 0.283369 +time/saving (s) 0.00411226 +time/training (s) 7.88924 +time/epoch (s) 10.8678 +time/total (s) 9351.92 +Epoch -98 +---------------------------------- --------------- +2022-05-10 15:46:53.153811 PDT | [1] Epoch -97 finished +---------------------------------- --------------- +epoch -97 +replay_buffer/size 999033 +trainer/num train calls 904000 +trainer/Policy Loss -19.562 +trainer/Log Pis Mean 24.9642 +trainer/Log Pis Std 13.7383 +trainer/Log Pis Max 74.4153 +trainer/Log Pis Min -5.96199 +trainer/policy/mean Mean -0.0184625 +trainer/policy/mean Std 0.909557 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80433 +trainer/policy/normal/std Std 0.681479 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.278288 +trainer/policy/normal/log_std Mean 0.987754 +trainer/policy/normal/log_std Std 0.335135 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.2791 +eval/num steps total 901446 +eval/num paths total 904 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0556101 +eval/Actions Std 0.908677 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65767 +time/logging (s) 0.00373614 +time/sampling batch (s) 0.280185 +time/saving (s) 0.00356643 +time/training (s) 7.35936 +time/epoch (s) 10.3045 +time/total (s) 9362.22 +Epoch -97 +---------------------------------- --------------- +2022-05-10 15:47:04.078301 PDT | [1] Epoch -96 finished +---------------------------------- --------------- +epoch -96 +replay_buffer/size 999033 +trainer/num train calls 905000 +trainer/Policy Loss -18.8198 +trainer/Log Pis Mean 23.7323 +trainer/Log Pis Std 12.8074 +trainer/Log Pis Max 61.1054 +trainer/Log Pis Min -9.88273 +trainer/policy/mean Mean -0.0334894 +trainer/policy/mean Std 0.903015 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79419 +trainer/policy/normal/std Std 0.682223 +trainer/policy/normal/std Max 5.63419 +trainer/policy/normal/std Min 0.256985 +trainer/policy/normal/log_std Mean 0.983559 +trainer/policy/normal/log_std Std 0.338282 +trainer/policy/normal/log_std Max 1.72885 +trainer/policy/normal/log_std Min -1.35874 +eval/num steps total 902446 +eval/num paths total 905 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104194 +eval/Actions Std 0.906352 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70075 +time/logging (s) 0.0037 +time/sampling batch (s) 0.530199 +time/saving (s) 0.00342971 +time/training (s) 7.66121 +time/epoch (s) 10.8993 +time/total (s) 9373.13 +Epoch -96 +---------------------------------- --------------- +2022-05-10 15:47:14.423218 PDT | [1] Epoch -95 finished +---------------------------------- --------------- +epoch -95 +replay_buffer/size 999033 +trainer/num train calls 906000 +trainer/Policy Loss -19.1394 +trainer/Log Pis Mean 24.4894 +trainer/Log Pis Std 13.0273 +trainer/Log Pis Max 66.845 +trainer/Log Pis Min -7.46943 +trainer/policy/mean Mean -0.0289808 +trainer/policy/mean Std 0.903617 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82647 +trainer/policy/normal/std Std 0.667731 +trainer/policy/normal/std Max 5.45467 +trainer/policy/normal/std Min 0.297515 +trainer/policy/normal/log_std Mean 0.997947 +trainer/policy/normal/log_std Std 0.325406 +trainer/policy/normal/log_std Max 1.69647 +trainer/policy/normal/log_std Min -1.21229 +eval/num steps total 903446 +eval/num paths total 906 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0829386 +eval/Actions Std 0.913473 +eval/Actions Max 0.999978 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60288 +time/logging (s) 0.00381269 +time/sampling batch (s) 0.27704 +time/saving (s) 0.00341545 +time/training (s) 7.43301 +time/epoch (s) 10.3202 +time/total (s) 9383.45 +Epoch -95 +---------------------------------- --------------- +2022-05-10 15:47:24.736468 PDT | [1] Epoch -94 finished +---------------------------------- --------------- +epoch -94 +replay_buffer/size 999033 +trainer/num train calls 907000 +trainer/Policy Loss -20.5056 +trainer/Log Pis Mean 25.0253 +trainer/Log Pis Std 13.6031 +trainer/Log Pis Max 63.9755 +trainer/Log Pis Min -5.73083 +trainer/policy/mean Mean -0.0537255 +trainer/policy/mean Std 0.906583 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77902 +trainer/policy/normal/std Std 0.670874 +trainer/policy/normal/std Max 5.47491 +trainer/policy/normal/std Min 0.239544 +trainer/policy/normal/log_std Mean 0.979521 +trainer/policy/normal/log_std Std 0.33047 +trainer/policy/normal/log_std Max 1.70018 +trainer/policy/normal/log_std Min -1.42902 +eval/num steps total 904446 +eval/num paths total 907 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100881 +eval/Actions Std 0.916881 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49285 +time/logging (s) 0.003755 +time/sampling batch (s) 0.526809 +time/saving (s) 0.00346546 +time/training (s) 7.26127 +time/epoch (s) 10.2881 +time/total (s) 9393.74 +Epoch -94 +---------------------------------- --------------- +2022-05-10 15:47:36.097653 PDT | [1] Epoch -93 finished +---------------------------------- --------------- +epoch -93 +replay_buffer/size 999033 +trainer/num train calls 908000 +trainer/Policy Loss -20.1394 +trainer/Log Pis Mean 24.6057 +trainer/Log Pis Std 14.3208 +trainer/Log Pis Max 71.3316 +trainer/Log Pis Min -9.16783 +trainer/policy/mean Mean -0.0334941 +trainer/policy/mean Std 0.906215 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79168 +trainer/policy/normal/std Std 0.683244 +trainer/policy/normal/std Max 6.93191 +trainer/policy/normal/std Min 0.27527 +trainer/policy/normal/log_std Mean 0.982266 +trainer/policy/normal/log_std Std 0.340295 +trainer/policy/normal/log_std Max 1.93614 +trainer/policy/normal/log_std Min -1.29 +eval/num steps total 905446 +eval/num paths total 908 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0167126 +eval/Actions Std 0.871344 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49533 +time/logging (s) 0.00369793 +time/sampling batch (s) 1.02873 +time/saving (s) 0.00341143 +time/training (s) 7.80497 +time/epoch (s) 11.3361 +time/total (s) 9405.08 +Epoch -93 +---------------------------------- --------------- +2022-05-10 15:47:47.349617 PDT | [1] Epoch -92 finished +---------------------------------- --------------- +epoch -92 +replay_buffer/size 999033 +trainer/num train calls 909000 +trainer/Policy Loss -21.3899 +trainer/Log Pis Mean 24.7354 +trainer/Log Pis Std 13.9829 +trainer/Log Pis Max 71.948 +trainer/Log Pis Min -7.69373 +trainer/policy/mean Mean -0.0278494 +trainer/policy/mean Std 0.907338 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79959 +trainer/policy/normal/std Std 0.692515 +trainer/policy/normal/std Max 5.39492 +trainer/policy/normal/std Min 0.253297 +trainer/policy/normal/log_std Mean 0.983476 +trainer/policy/normal/log_std Std 0.347227 +trainer/policy/normal/log_std Max 1.68546 +trainer/policy/normal/log_std Min -1.37319 +eval/num steps total 906446 +eval/num paths total 909 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0064898 +eval/Actions Std 0.903028 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51748 +time/logging (s) 0.00370806 +time/sampling batch (s) 1.02779 +time/saving (s) 0.00341697 +time/training (s) 7.6746 +time/epoch (s) 11.227 +time/total (s) 9416.31 +Epoch -92 +---------------------------------- --------------- +2022-05-10 15:47:58.355642 PDT | [1] Epoch -91 finished +---------------------------------- --------------- +epoch -91 +replay_buffer/size 999033 +trainer/num train calls 910000 +trainer/Policy Loss -19.4148 +trainer/Log Pis Mean 25.3338 +trainer/Log Pis Std 13.0285 +trainer/Log Pis Max 76.2928 +trainer/Log Pis Min -7.89516 +trainer/policy/mean Mean -0.0314205 +trainer/policy/mean Std 0.90912 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999975 +trainer/policy/normal/std Mean 2.84017 +trainer/policy/normal/std Std 0.664824 +trainer/policy/normal/std Max 5.05897 +trainer/policy/normal/std Min 0.285384 +trainer/policy/normal/log_std Mean 1.00415 +trainer/policy/normal/log_std Std 0.318123 +trainer/policy/normal/log_std Max 1.62116 +trainer/policy/normal/log_std Min -1.25392 +eval/num steps total 907446 +eval/num paths total 910 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0855884 +eval/Actions Std 0.904116 +eval/Actions Max 0.999997 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56687 +time/logging (s) 0.00385042 +time/sampling batch (s) 0.282497 +time/saving (s) 0.00390593 +time/training (s) 8.12376 +time/epoch (s) 10.9809 +time/total (s) 9427.29 +Epoch -91 +---------------------------------- --------------- +2022-05-10 15:48:08.481089 PDT | [1] Epoch -90 finished +---------------------------------- --------------- +epoch -90 +replay_buffer/size 999033 +trainer/num train calls 911000 +trainer/Policy Loss -19.3962 +trainer/Log Pis Mean 23.0358 +trainer/Log Pis Std 13.5879 +trainer/Log Pis Max 71.5949 +trainer/Log Pis Min -11.9948 +trainer/policy/mean Mean -0.0302212 +trainer/policy/mean Std 0.900555 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78124 +trainer/policy/normal/std Std 0.668539 +trainer/policy/normal/std Max 5.6184 +trainer/policy/normal/std Min 0.250224 +trainer/policy/normal/log_std Mean 0.97943 +trainer/policy/normal/log_std Std 0.337549 +trainer/policy/normal/log_std Max 1.72605 +trainer/policy/normal/log_std Min -1.3854 +eval/num steps total 908446 +eval/num paths total 911 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.381169 +eval/Actions Std 0.732742 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49058 +time/logging (s) 0.00421642 +time/sampling batch (s) 0.279581 +time/saving (s) 0.00381688 +time/training (s) 7.31642 +time/epoch (s) 10.0946 +time/total (s) 9437.39 +Epoch -90 +---------------------------------- --------------- +2022-05-10 15:48:17.988185 PDT | [1] Epoch -89 finished +---------------------------------- --------------- +epoch -89 +replay_buffer/size 999033 +trainer/num train calls 912000 +trainer/Policy Loss -19.8323 +trainer/Log Pis Mean 24.4567 +trainer/Log Pis Std 14.1029 +trainer/Log Pis Max 74.5928 +trainer/Log Pis Min -6.89058 +trainer/policy/mean Mean -0.0318002 +trainer/policy/mean Std 0.906099 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7909 +trainer/policy/normal/std Std 0.711265 +trainer/policy/normal/std Max 5.82131 +trainer/policy/normal/std Min 0.21672 +trainer/policy/normal/log_std Mean 0.978364 +trainer/policy/normal/log_std Std 0.354444 +trainer/policy/normal/log_std Max 1.76153 +trainer/policy/normal/log_std Min -1.52915 +eval/num steps total 909446 +eval/num paths total 912 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.210973 +eval/Actions Std 0.901951 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44981 +time/logging (s) 0.00391622 +time/sampling batch (s) 0.529549 +time/saving (s) 0.0036237 +time/training (s) 6.49416 +time/epoch (s) 9.48106 +time/total (s) 9446.88 +Epoch -89 +---------------------------------- --------------- +2022-05-10 15:48:27.513513 PDT | [1] Epoch -88 finished +---------------------------------- --------------- +epoch -88 +replay_buffer/size 999033 +trainer/num train calls 913000 +trainer/Policy Loss -21.2023 +trainer/Log Pis Mean 24.2026 +trainer/Log Pis Std 13.1041 +trainer/Log Pis Max 73.8161 +trainer/Log Pis Min -8.11127 +trainer/policy/mean Mean -0.0479118 +trainer/policy/mean Std 0.907864 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7953 +trainer/policy/normal/std Std 0.69417 +trainer/policy/normal/std Max 5.80209 +trainer/policy/normal/std Min 0.229135 +trainer/policy/normal/log_std Mean 0.981602 +trainer/policy/normal/log_std Std 0.347995 +trainer/policy/normal/log_std Max 1.75822 +trainer/policy/normal/log_std Min -1.47344 +eval/num steps total 910446 +eval/num paths total 913 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101321 +eval/Actions Std 0.919887 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70415 +time/logging (s) 0.0039076 +time/sampling batch (s) 0.280651 +time/saving (s) 0.00355601 +time/training (s) 6.50781 +time/epoch (s) 9.50008 +time/total (s) 9456.38 +Epoch -88 +---------------------------------- --------------- +2022-05-10 15:48:37.248703 PDT | [1] Epoch -87 finished +---------------------------------- --------------- +epoch -87 +replay_buffer/size 999033 +trainer/num train calls 914000 +trainer/Policy Loss -20.2415 +trainer/Log Pis Mean 24.3967 +trainer/Log Pis Std 14.1121 +trainer/Log Pis Max 72.8275 +trainer/Log Pis Min -7.51339 +trainer/policy/mean Mean -0.0457526 +trainer/policy/mean Std 0.908312 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79073 +trainer/policy/normal/std Std 0.690626 +trainer/policy/normal/std Max 6.06605 +trainer/policy/normal/std Min 0.25848 +trainer/policy/normal/log_std Mean 0.981191 +trainer/policy/normal/log_std Std 0.341737 +trainer/policy/normal/log_std Max 1.80271 +trainer/policy/normal/log_std Min -1.35294 +eval/num steps total 911446 +eval/num paths total 914 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.211476 +eval/Actions Std 0.925082 +eval/Actions Max 0.999987 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54656 +time/logging (s) 0.00376044 +time/sampling batch (s) 0.528215 +time/saving (s) 0.00345476 +time/training (s) 6.62783 +time/epoch (s) 9.70982 +time/total (s) 9466.09 +Epoch -87 +---------------------------------- --------------- +2022-05-10 15:48:48.242989 PDT | [1] Epoch -86 finished +---------------------------------- --------------- +epoch -86 +replay_buffer/size 999033 +trainer/num train calls 915000 +trainer/Policy Loss -19.7304 +trainer/Log Pis Mean 24.4897 +trainer/Log Pis Std 13.1627 +trainer/Log Pis Max 68.2375 +trainer/Log Pis Min -6.00346 +trainer/policy/mean Mean -0.0115395 +trainer/policy/mean Std 0.903461 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79553 +trainer/policy/normal/std Std 0.693696 +trainer/policy/normal/std Max 5.76723 +trainer/policy/normal/std Min 0.220424 +trainer/policy/normal/log_std Mean 0.980875 +trainer/policy/normal/log_std Std 0.353535 +trainer/policy/normal/log_std Max 1.75219 +trainer/policy/normal/log_std Min -1.5122 +eval/num steps total 912446 +eval/num paths total 915 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0747278 +eval/Actions Std 0.915514 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59601 +time/logging (s) 0.00377573 +time/sampling batch (s) 0.279108 +time/saving (s) 0.00351774 +time/training (s) 8.08676 +time/epoch (s) 10.9692 +time/total (s) 9477.06 +Epoch -86 +---------------------------------- --------------- +2022-05-10 15:48:59.426186 PDT | [1] Epoch -85 finished +---------------------------------- --------------- +epoch -85 +replay_buffer/size 999033 +trainer/num train calls 916000 +trainer/Policy Loss -19.9695 +trainer/Log Pis Mean 24.3069 +trainer/Log Pis Std 13.5361 +trainer/Log Pis Max 73.9015 +trainer/Log Pis Min -8.65245 +trainer/policy/mean Mean -0.0331888 +trainer/policy/mean Std 0.912161 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79054 +trainer/policy/normal/std Std 0.638225 +trainer/policy/normal/std Max 5.96575 +trainer/policy/normal/std Min 0.224238 +trainer/policy/normal/log_std Mean 0.989649 +trainer/policy/normal/log_std Std 0.304787 +trainer/policy/normal/log_std Max 1.78604 +trainer/policy/normal/log_std Min -1.49505 +eval/num steps total 913446 +eval/num paths total 916 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200685 +eval/Actions Std 0.891727 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75897 +time/logging (s) 0.00378013 +time/sampling batch (s) 0.528126 +time/saving (s) 0.00347123 +time/training (s) 7.86383 +time/epoch (s) 11.1582 +time/total (s) 9488.23 +Epoch -85 +---------------------------------- --------------- +2022-05-10 15:49:09.576151 PDT | [1] Epoch -84 finished +---------------------------------- --------------- +epoch -84 +replay_buffer/size 999033 +trainer/num train calls 917000 +trainer/Policy Loss -20.2836 +trainer/Log Pis Mean 25.3953 +trainer/Log Pis Std 13.3673 +trainer/Log Pis Max 63.2629 +trainer/Log Pis Min -4.45668 +trainer/policy/mean Mean -0.0364244 +trainer/policy/mean Std 0.907966 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82424 +trainer/policy/normal/std Std 0.681142 +trainer/policy/normal/std Max 6.17449 +trainer/policy/normal/std Min 0.332028 +trainer/policy/normal/log_std Mean 0.995506 +trainer/policy/normal/log_std Std 0.332072 +trainer/policy/normal/log_std Max 1.82043 +trainer/policy/normal/log_std Min -1.10254 +eval/num steps total 914446 +eval/num paths total 917 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.270822 +eval/Actions Std 0.888588 +eval/Actions Max 0.999985 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57379 +time/logging (s) 0.00388622 +time/sampling batch (s) 0.275125 +time/saving (s) 0.00342233 +time/training (s) 7.26894 +time/epoch (s) 10.1252 +time/total (s) 9498.35 +Epoch -84 +---------------------------------- --------------- +2022-05-10 15:49:19.680787 PDT | [1] Epoch -83 finished +---------------------------------- --------------- +epoch -83 +replay_buffer/size 999033 +trainer/num train calls 918000 +trainer/Policy Loss -20.2884 +trainer/Log Pis Mean 23.8984 +trainer/Log Pis Std 13.3359 +trainer/Log Pis Max 66.1079 +trainer/Log Pis Min -11.251 +trainer/policy/mean Mean -0.0379258 +trainer/policy/mean Std 0.903591 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80433 +trainer/policy/normal/std Std 0.709531 +trainer/policy/normal/std Max 5.96869 +trainer/policy/normal/std Min 0.299037 +trainer/policy/normal/log_std Mean 0.983379 +trainer/policy/normal/log_std Std 0.352532 +trainer/policy/normal/log_std Max 1.78653 +trainer/policy/normal/log_std Min -1.20719 +eval/num steps total 915446 +eval/num paths total 918 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.421071 +eval/Actions Std 0.858494 +eval/Actions Max 0.999997 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37092 +time/logging (s) 0.00397749 +time/sampling batch (s) 0.274253 +time/saving (s) 0.00377375 +time/training (s) 7.4271 +time/epoch (s) 10.08 +time/total (s) 9508.44 +Epoch -83 +---------------------------------- --------------- +2022-05-10 15:49:28.968284 PDT | [1] Epoch -82 finished +---------------------------------- --------------- +epoch -82 +replay_buffer/size 999033 +trainer/num train calls 919000 +trainer/Policy Loss -19.8467 +trainer/Log Pis Mean 23.7861 +trainer/Log Pis Std 13.0214 +trainer/Log Pis Max 80.27 +trainer/Log Pis Min -8.22858 +trainer/policy/mean Mean -0.0380318 +trainer/policy/mean Std 0.90912 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.75226 +trainer/policy/normal/std Std 0.669708 +trainer/policy/normal/std Max 6.3949 +trainer/policy/normal/std Min 0.229934 +trainer/policy/normal/log_std Mean 0.968788 +trainer/policy/normal/log_std Std 0.337511 +trainer/policy/normal/log_std Max 1.8555 +trainer/policy/normal/log_std Min -1.46996 +eval/num steps total 916446 +eval/num paths total 919 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.073354 +eval/Actions Std 0.905179 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58025 +time/logging (s) 0.00386365 +time/sampling batch (s) 0.275371 +time/saving (s) 0.00352239 +time/training (s) 6.39928 +time/epoch (s) 9.26229 +time/total (s) 9517.7 +Epoch -82 +---------------------------------- --------------- +2022-05-10 15:49:39.278848 PDT | [1] Epoch -81 finished +---------------------------------- --------------- +epoch -81 +replay_buffer/size 999033 +trainer/num train calls 920000 +trainer/Policy Loss -19.8635 +trainer/Log Pis Mean 24.1576 +trainer/Log Pis Std 13.4361 +trainer/Log Pis Max 60.4218 +trainer/Log Pis Min -6.60729 +trainer/policy/mean Mean -0.0355809 +trainer/policy/mean Std 0.904524 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86741 +trainer/policy/normal/std Std 0.709709 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.331206 +trainer/policy/normal/log_std Mean 1.00795 +trainer/policy/normal/log_std Std 0.343801 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.10502 +eval/num steps total 917446 +eval/num paths total 920 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.281402 +eval/Actions Std 0.879281 +eval/Actions Max 0.999989 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72034 +time/logging (s) 0.00374306 +time/sampling batch (s) 0.275095 +time/saving (s) 0.00340285 +time/training (s) 7.28318 +time/epoch (s) 10.2858 +time/total (s) 9527.99 +Epoch -81 +---------------------------------- --------------- +2022-05-10 15:49:51.336838 PDT | [1] Epoch -80 finished +---------------------------------- --------------- +epoch -80 +replay_buffer/size 999033 +trainer/num train calls 921000 +trainer/Policy Loss -18.8724 +trainer/Log Pis Mean 24.3663 +trainer/Log Pis Std 13.4798 +trainer/Log Pis Max 69.1818 +trainer/Log Pis Min -7.35769 +trainer/policy/mean Mean -0.0313849 +trainer/policy/mean Std 0.905154 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78751 +trainer/policy/normal/std Std 0.704522 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.27357 +trainer/policy/normal/log_std Mean 0.978329 +trainer/policy/normal/log_std Std 0.347986 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.2962 +eval/num steps total 918446 +eval/num paths total 921 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.282141 +eval/Actions Std 0.80589 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66325 +time/logging (s) 0.00383612 +time/sampling batch (s) 1.03615 +time/saving (s) 0.00348502 +time/training (s) 8.32605 +time/epoch (s) 12.0328 +time/total (s) 9540.03 +Epoch -80 +---------------------------------- --------------- +2022-05-10 15:50:01.192176 PDT | [1] Epoch -79 finished +---------------------------------- --------------- +epoch -79 +replay_buffer/size 999033 +trainer/num train calls 922000 +trainer/Policy Loss -19.8901 +trainer/Log Pis Mean 25.2193 +trainer/Log Pis Std 12.9237 +trainer/Log Pis Max 76.2762 +trainer/Log Pis Min -4.67692 +trainer/policy/mean Mean -0.0560518 +trainer/policy/mean Std 0.908997 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83765 +trainer/policy/normal/std Std 0.660149 +trainer/policy/normal/std Max 6.10588 +trainer/policy/normal/std Min 0.289599 +trainer/policy/normal/log_std Mean 1.00397 +trainer/policy/normal/log_std Std 0.316158 +trainer/policy/normal/log_std Max 1.80925 +trainer/policy/normal/log_std Min -1.23926 +eval/num steps total 919446 +eval/num paths total 922 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.189285 +eval/Actions Std 0.914124 +eval/Actions Max 0.999994 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6844 +time/logging (s) 0.00383474 +time/sampling batch (s) 0.280538 +time/saving (s) 0.0035002 +time/training (s) 6.85784 +time/epoch (s) 9.83011 +time/total (s) 9549.86 +Epoch -79 +---------------------------------- --------------- +2022-05-10 15:50:11.395346 PDT | [1] Epoch -78 finished +---------------------------------- --------------- +epoch -78 +replay_buffer/size 999033 +trainer/num train calls 923000 +trainer/Policy Loss -19.8283 +trainer/Log Pis Mean 23.4792 +trainer/Log Pis Std 12.7326 +trainer/Log Pis Max 73.3956 +trainer/Log Pis Min -6.67302 +trainer/policy/mean Mean -0.0219213 +trainer/policy/mean Std 0.908777 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76992 +trainer/policy/normal/std Std 0.686913 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.303455 +trainer/policy/normal/log_std Mean 0.973845 +trainer/policy/normal/log_std Std 0.340772 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19252 +eval/num steps total 920446 +eval/num paths total 923 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0118941 +eval/Actions Std 0.902291 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73371 +time/logging (s) 0.00420129 +time/sampling batch (s) 0.281972 +time/saving (s) 0.0040206 +time/training (s) 7.15382 +time/epoch (s) 10.1777 +time/total (s) 9560.04 +Epoch -78 +---------------------------------- --------------- +2022-05-10 15:50:22.254360 PDT | [1] Epoch -77 finished +---------------------------------- --------------- +epoch -77 +replay_buffer/size 999033 +trainer/num train calls 924000 +trainer/Policy Loss -19.6267 +trainer/Log Pis Mean 24.5353 +trainer/Log Pis Std 13.0113 +trainer/Log Pis Max 66.2197 +trainer/Log Pis Min -7.68258 +trainer/policy/mean Mean -0.0345656 +trainer/policy/mean Std 0.909276 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79715 +trainer/policy/normal/std Std 0.673173 +trainer/policy/normal/std Max 5.54378 +trainer/policy/normal/std Min 0.230946 +trainer/policy/normal/log_std Mean 0.986827 +trainer/policy/normal/log_std Std 0.326732 +trainer/policy/normal/log_std Max 1.71268 +trainer/policy/normal/log_std Min -1.46557 +eval/num steps total 921446 +eval/num paths total 924 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116669 +eval/Actions Std 0.89882 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71803 +time/logging (s) 0.00389122 +time/sampling batch (s) 0.280988 +time/saving (s) 0.00374688 +time/training (s) 7.82603 +time/epoch (s) 10.8327 +time/total (s) 9570.88 +Epoch -77 +---------------------------------- --------------- +2022-05-10 15:50:32.916887 PDT | [1] Epoch -76 finished +---------------------------------- --------------- +epoch -76 +replay_buffer/size 999033 +trainer/num train calls 925000 +trainer/Policy Loss -19.1669 +trainer/Log Pis Mean 24.884 +trainer/Log Pis Std 13.8028 +trainer/Log Pis Max 71.2871 +trainer/Log Pis Min -10.1153 +trainer/policy/mean Mean -0.0401677 +trainer/policy/mean Std 0.903112 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84012 +trainer/policy/normal/std Std 0.691753 +trainer/policy/normal/std Max 6.2111 +trainer/policy/normal/std Min 0.286023 +trainer/policy/normal/log_std Mean 0.999687 +trainer/policy/normal/log_std Std 0.339079 +trainer/policy/normal/log_std Max 1.82634 +trainer/policy/normal/log_std Min -1.25168 +eval/num steps total 922446 +eval/num paths total 925 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0825394 +eval/Actions Std 0.896834 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60898 +time/logging (s) 0.00369704 +time/sampling batch (s) 0.303301 +time/saving (s) 0.00344143 +time/training (s) 7.71725 +time/epoch (s) 10.6367 +time/total (s) 9581.52 +Epoch -76 +---------------------------------- --------------- +2022-05-10 15:50:44.643952 PDT | [1] Epoch -75 finished +---------------------------------- --------------- +epoch -75 +replay_buffer/size 999033 +trainer/num train calls 926000 +trainer/Policy Loss -18.9232 +trainer/Log Pis Mean 25.3615 +trainer/Log Pis Std 14.365 +trainer/Log Pis Max 81.8588 +trainer/Log Pis Min -10.4495 +trainer/policy/mean Mean -0.0492956 +trainer/policy/mean Std 0.910038 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86199 +trainer/policy/normal/std Std 0.683343 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.345815 +trainer/policy/normal/log_std Mean 1.0097 +trainer/policy/normal/log_std Std 0.329109 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.06185 +eval/num steps total 923446 +eval/num paths total 926 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.153067 +eval/Actions Std 0.88247 +eval/Actions Max 0.999975 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55479 +time/logging (s) 0.0037822 +time/sampling batch (s) 0.531832 +time/saving (s) 0.00349386 +time/training (s) 8.60786 +time/epoch (s) 11.7018 +time/total (s) 9593.22 +Epoch -75 +---------------------------------- --------------- +2022-05-10 15:50:54.599327 PDT | [1] Epoch -74 finished +---------------------------------- -------------- +epoch -74 +replay_buffer/size 999033 +trainer/num train calls 927000 +trainer/Policy Loss -20.2469 +trainer/Log Pis Mean 24.7251 +trainer/Log Pis Std 14.0301 +trainer/Log Pis Max 83.9809 +trainer/Log Pis Min -8.44334 +trainer/policy/mean Mean -0.0513661 +trainer/policy/mean Std 0.90956 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77975 +trainer/policy/normal/std Std 0.668586 +trainer/policy/normal/std Max 5.80293 +trainer/policy/normal/std Min 0.278284 +trainer/policy/normal/log_std Mean 0.980811 +trainer/policy/normal/log_std Std 0.32535 +trainer/policy/normal/log_std Max 1.75836 +trainer/policy/normal/log_std Min -1.27911 +eval/num steps total 924446 +eval/num paths total 927 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129522 +eval/Actions Std 0.905999 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64619 +time/logging (s) 0.0038607 +time/sampling batch (s) 0.280082 +time/saving (s) 0.0034091 +time/training (s) 6.99633 +time/epoch (s) 9.92987 +time/total (s) 9603.15 +Epoch -74 +---------------------------------- -------------- +2022-05-10 15:51:05.517368 PDT | [1] Epoch -73 finished +---------------------------------- --------------- +epoch -73 +replay_buffer/size 999033 +trainer/num train calls 928000 +trainer/Policy Loss -20.1345 +trainer/Log Pis Mean 25.3682 +trainer/Log Pis Std 13.7867 +trainer/Log Pis Max 69.0212 +trainer/Log Pis Min -8.6591 +trainer/policy/mean Mean -0.0513204 +trainer/policy/mean Std 0.907891 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79455 +trainer/policy/normal/std Std 0.667489 +trainer/policy/normal/std Max 5.36539 +trainer/policy/normal/std Min 0.196369 +trainer/policy/normal/log_std Mean 0.983253 +trainer/policy/normal/log_std Std 0.344575 +trainer/policy/normal/log_std Max 1.67997 +trainer/policy/normal/log_std Min -1.62776 +eval/num steps total 925446 +eval/num paths total 928 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0693059 +eval/Actions Std 0.912585 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67997 +time/logging (s) 0.00411682 +time/sampling batch (s) 0.281687 +time/saving (s) 0.00385986 +time/training (s) 7.92305 +time/epoch (s) 10.8927 +time/total (s) 9614.05 +Epoch -73 +---------------------------------- --------------- +2022-05-10 15:51:17.555481 PDT | [1] Epoch -72 finished +---------------------------------- --------------- +epoch -72 +replay_buffer/size 999033 +trainer/num train calls 929000 +trainer/Policy Loss -19.9075 +trainer/Log Pis Mean 24.455 +trainer/Log Pis Std 13.2317 +trainer/Log Pis Max 73.8771 +trainer/Log Pis Min -6.50351 +trainer/policy/mean Mean -0.0276892 +trainer/policy/mean Std 0.906165 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82191 +trainer/policy/normal/std Std 0.700039 +trainer/policy/normal/std Max 5.83345 +trainer/policy/normal/std Min 0.242808 +trainer/policy/normal/log_std Mean 0.990211 +trainer/policy/normal/log_std Std 0.35345 +trainer/policy/normal/log_std Max 1.76361 +trainer/policy/normal/log_std Min -1.41549 +eval/num steps total 926446 +eval/num paths total 929 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0536715 +eval/Actions Std 0.895334 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73587 +time/logging (s) 0.00383576 +time/sampling batch (s) 0.281765 +time/saving (s) 0.00406599 +time/training (s) 8.98572 +time/epoch (s) 12.0113 +time/total (s) 9626.07 +Epoch -72 +---------------------------------- --------------- +2022-05-10 15:51:27.566059 PDT | [1] Epoch -71 finished +---------------------------------- --------------- +epoch -71 +replay_buffer/size 999033 +trainer/num train calls 930000 +trainer/Policy Loss -21.2917 +trainer/Log Pis Mean 25.3565 +trainer/Log Pis Std 14.0495 +trainer/Log Pis Max 77.0106 +trainer/Log Pis Min -9.817 +trainer/policy/mean Mean -0.0581483 +trainer/policy/mean Std 0.911511 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.823 +trainer/policy/normal/std Std 0.64735 +trainer/policy/normal/std Max 5.46894 +trainer/policy/normal/std Min 0.29426 +trainer/policy/normal/log_std Mean 0.999119 +trainer/policy/normal/log_std Std 0.316454 +trainer/policy/normal/log_std Max 1.69908 +trainer/policy/normal/log_std Min -1.22329 +eval/num steps total 927446 +eval/num paths total 930 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116279 +eval/Actions Std 0.895475 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.94882 +time/logging (s) 0.00369459 +time/sampling batch (s) 0.28176 +time/saving (s) 0.00346727 +time/training (s) 6.74605 +time/epoch (s) 9.98379 +time/total (s) 9636.05 +Epoch -71 +---------------------------------- --------------- +2022-05-10 15:51:37.711042 PDT | [1] Epoch -70 finished +---------------------------------- --------------- +epoch -70 +replay_buffer/size 999033 +trainer/num train calls 931000 +trainer/Policy Loss -19.5145 +trainer/Log Pis Mean 23.6013 +trainer/Log Pis Std 13.06 +trainer/Log Pis Max 66.7081 +trainer/Log Pis Min -8.4204 +trainer/policy/mean Mean -0.0428011 +trainer/policy/mean Std 0.903272 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80776 +trainer/policy/normal/std Std 0.658303 +trainer/policy/normal/std Max 6.85202 +trainer/policy/normal/std Min 0.341588 +trainer/policy/normal/log_std Mean 0.993168 +trainer/policy/normal/log_std Std 0.315554 +trainer/policy/normal/log_std Max 1.92454 +trainer/policy/normal/log_std Min -1.07415 +eval/num steps total 928446 +eval/num paths total 931 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.206151 +eval/Actions Std 0.915555 +eval/Actions Max 0.999974 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83673 +time/logging (s) 0.00371779 +time/sampling batch (s) 0.531934 +time/saving (s) 0.00345697 +time/training (s) 6.74376 +time/epoch (s) 10.1196 +time/total (s) 9646.17 +Epoch -70 +---------------------------------- --------------- +2022-05-10 15:51:47.086383 PDT | [1] Epoch -69 finished +---------------------------------- --------------- +epoch -69 +replay_buffer/size 999033 +trainer/num train calls 932000 +trainer/Policy Loss -19.8437 +trainer/Log Pis Mean 24.9111 +trainer/Log Pis Std 13.6696 +trainer/Log Pis Max 81.8419 +trainer/Log Pis Min -10.7651 +trainer/policy/mean Mean -0.0417985 +trainer/policy/mean Std 0.910178 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.83516 +trainer/policy/normal/std Std 0.675664 +trainer/policy/normal/std Max 5.66026 +trainer/policy/normal/std Min 0.22976 +trainer/policy/normal/log_std Mean 1.00048 +trainer/policy/normal/log_std Std 0.328822 +trainer/policy/normal/log_std Max 1.73347 +trainer/policy/normal/log_std Min -1.47072 +eval/num steps total 929446 +eval/num paths total 932 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.485425 +eval/Actions Std 0.869057 +eval/Actions Max 0.999991 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69104 +time/logging (s) 0.00379698 +time/sampling batch (s) 0.279098 +time/saving (s) 0.00345018 +time/training (s) 6.37282 +time/epoch (s) 9.3502 +time/total (s) 9655.53 +Epoch -69 +---------------------------------- --------------- +2022-05-10 15:51:58.057903 PDT | [1] Epoch -68 finished +---------------------------------- --------------- +epoch -68 +replay_buffer/size 999033 +trainer/num train calls 933000 +trainer/Policy Loss -20.9829 +trainer/Log Pis Mean 25.1963 +trainer/Log Pis Std 13.2635 +trainer/Log Pis Max 69.4016 +trainer/Log Pis Min -8.28242 +trainer/policy/mean Mean -0.0575124 +trainer/policy/mean Std 0.907764 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77172 +trainer/policy/normal/std Std 0.66277 +trainer/policy/normal/std Max 5.56639 +trainer/policy/normal/std Min 0.204341 +trainer/policy/normal/log_std Mean 0.977659 +trainer/policy/normal/log_std Std 0.328559 +trainer/policy/normal/log_std Max 1.71675 +trainer/policy/normal/log_std Min -1.58797 +eval/num steps total 930446 +eval/num paths total 933 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.279004 +eval/Actions Std 0.816261 +eval/Actions Max 0.99999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8234 +time/logging (s) 0.00410621 +time/sampling batch (s) 0.284373 +time/saving (s) 0.00406427 +time/training (s) 7.83051 +time/epoch (s) 10.9464 +time/total (s) 9666.48 +Epoch -68 +---------------------------------- --------------- +2022-05-10 15:52:10.065502 PDT | [1] Epoch -67 finished +---------------------------------- --------------- +epoch -67 +replay_buffer/size 999033 +trainer/num train calls 934000 +trainer/Policy Loss -20.3341 +trainer/Log Pis Mean 24.5655 +trainer/Log Pis Std 14.4043 +trainer/Log Pis Max 77.5709 +trainer/Log Pis Min -8.40914 +trainer/policy/mean Mean -0.0159986 +trainer/policy/mean Std 0.908299 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82505 +trainer/policy/normal/std Std 0.673069 +trainer/policy/normal/std Max 5.97242 +trainer/policy/normal/std Min 0.292806 +trainer/policy/normal/log_std Mean 0.996418 +trainer/policy/normal/log_std Std 0.3303 +trainer/policy/normal/log_std Max 1.78715 +trainer/policy/normal/log_std Min -1.22825 +eval/num steps total 931446 +eval/num paths total 934 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.306064 +eval/Actions Std 0.851464 +eval/Actions Max 0.999994 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45263 +time/logging (s) 0.0037199 +time/sampling batch (s) 0.526348 +time/saving (s) 0.00340125 +time/training (s) 8.99561 +time/epoch (s) 11.9817 +time/total (s) 9678.46 +Epoch -67 +---------------------------------- --------------- +2022-05-10 15:52:20.072056 PDT | [1] Epoch -66 finished +---------------------------------- --------------- +epoch -66 +replay_buffer/size 999033 +trainer/num train calls 935000 +trainer/Policy Loss -20.0241 +trainer/Log Pis Mean 25.1637 +trainer/Log Pis Std 13.0254 +trainer/Log Pis Max 73.5597 +trainer/Log Pis Min -4.69258 +trainer/policy/mean Mean -0.0555645 +trainer/policy/mean Std 0.90524 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80114 +trainer/policy/normal/std Std 0.690726 +trainer/policy/normal/std Max 6.4758 +trainer/policy/normal/std Min 0.226934 +trainer/policy/normal/log_std Mean 0.985672 +trainer/policy/normal/log_std Std 0.337064 +trainer/policy/normal/log_std Max 1.86807 +trainer/policy/normal/log_std Min -1.4831 +eval/num steps total 932446 +eval/num paths total 935 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.164603 +eval/Actions Std 0.810472 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44342 +time/logging (s) 0.00379314 +time/sampling batch (s) 0.525465 +time/saving (s) 0.00345419 +time/training (s) 7.00553 +time/epoch (s) 9.98165 +time/total (s) 9688.45 +Epoch -66 +---------------------------------- --------------- +2022-05-10 15:52:30.321745 PDT | [1] Epoch -65 finished +---------------------------------- --------------- +epoch -65 +replay_buffer/size 999033 +trainer/num train calls 936000 +trainer/Policy Loss -19.5991 +trainer/Log Pis Mean 23.7671 +trainer/Log Pis Std 13.5844 +trainer/Log Pis Max 65.922 +trainer/Log Pis Min -7.6428 +trainer/policy/mean Mean -0.0594638 +trainer/policy/mean Std 0.906354 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.76244 +trainer/policy/normal/std Std 0.654156 +trainer/policy/normal/std Max 5.76889 +trainer/policy/normal/std Min 0.283751 +trainer/policy/normal/log_std Mean 0.97585 +trainer/policy/normal/log_std Std 0.319775 +trainer/policy/normal/log_std Max 1.75248 +trainer/policy/normal/log_std Min -1.25966 +eval/num steps total 933446 +eval/num paths total 936 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.308419 +eval/Actions Std 0.852379 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.686 +time/logging (s) 0.00372455 +time/sampling batch (s) 0.274469 +time/saving (s) 0.00345087 +time/training (s) 7.25699 +time/epoch (s) 10.2246 +time/total (s) 9698.67 +Epoch -65 +---------------------------------- --------------- +2022-05-10 15:52:40.764236 PDT | [1] Epoch -64 finished +---------------------------------- --------------- +epoch -64 +replay_buffer/size 999033 +trainer/num train calls 937000 +trainer/Policy Loss -19.7217 +trainer/Log Pis Mean 24.2795 +trainer/Log Pis Std 13.5312 +trainer/Log Pis Max 76.8217 +trainer/Log Pis Min -8.08811 +trainer/policy/mean Mean -0.00755623 +trainer/policy/mean Std 0.906211 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79736 +trainer/policy/normal/std Std 0.679575 +trainer/policy/normal/std Max 7.38373 +trainer/policy/normal/std Min 0.249801 +trainer/policy/normal/log_std Mean 0.985782 +trainer/policy/normal/log_std Std 0.331379 +trainer/policy/normal/log_std Max 1.99928 +trainer/policy/normal/log_std Min -1.38709 +eval/num steps total 934446 +eval/num paths total 937 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.427499 +eval/Actions Std 0.814279 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56681 +time/logging (s) 0.00486206 +time/sampling batch (s) 0.27833 +time/saving (s) 0.00527741 +time/training (s) 7.56317 +time/epoch (s) 10.4184 +time/total (s) 9709.1 +Epoch -64 +---------------------------------- --------------- +2022-05-10 15:52:51.103372 PDT | [1] Epoch -63 finished +---------------------------------- --------------- +epoch -63 +replay_buffer/size 999033 +trainer/num train calls 938000 +trainer/Policy Loss -20.5687 +trainer/Log Pis Mean 24.9537 +trainer/Log Pis Std 13.4757 +trainer/Log Pis Max 78.072 +trainer/Log Pis Min -6.20614 +trainer/policy/mean Mean -0.0329099 +trainer/policy/mean Std 0.906813 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83401 +trainer/policy/normal/std Std 0.698188 +trainer/policy/normal/std Max 6.76377 +trainer/policy/normal/std Min 0.298716 +trainer/policy/normal/log_std Mean 0.99658 +trainer/policy/normal/log_std Std 0.341647 +trainer/policy/normal/log_std Max 1.91158 +trainer/policy/normal/log_std Min -1.20826 +eval/num steps total 935446 +eval/num paths total 938 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.055137 +eval/Actions Std 0.903646 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59504 +time/logging (s) 0.00410168 +time/sampling batch (s) 0.277385 +time/saving (s) 0.00380954 +time/training (s) 7.43159 +time/epoch (s) 10.3119 +time/total (s) 9719.41 +Epoch -63 +---------------------------------- --------------- +2022-05-10 15:53:01.619022 PDT | [1] Epoch -62 finished +---------------------------------- --------------- +epoch -62 +replay_buffer/size 999033 +trainer/num train calls 939000 +trainer/Policy Loss -18.9924 +trainer/Log Pis Mean 23.2946 +trainer/Log Pis Std 12.792 +trainer/Log Pis Max 67.2752 +trainer/Log Pis Min -12.7719 +trainer/policy/mean Mean -0.0449155 +trainer/policy/mean Std 0.90528 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.74206 +trainer/policy/normal/std Std 0.660794 +trainer/policy/normal/std Max 6.77795 +trainer/policy/normal/std Min 0.237149 +trainer/policy/normal/log_std Mean 0.965279 +trainer/policy/normal/log_std Std 0.336094 +trainer/policy/normal/log_std Max 1.91367 +trainer/policy/normal/log_std Min -1.43907 +eval/num steps total 936446 +eval/num paths total 939 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0147607 +eval/Actions Std 0.905362 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56691 +time/logging (s) 0.00382545 +time/sampling batch (s) 0.281022 +time/saving (s) 0.00371942 +time/training (s) 7.63442 +time/epoch (s) 10.4899 +time/total (s) 9729.9 +Epoch -62 +---------------------------------- --------------- +2022-05-10 15:53:12.724591 PDT | [1] Epoch -61 finished +---------------------------------- --------------- +epoch -61 +replay_buffer/size 999033 +trainer/num train calls 940000 +trainer/Policy Loss -21.0818 +trainer/Log Pis Mean 24.3913 +trainer/Log Pis Std 14.3152 +trainer/Log Pis Max 76.841 +trainer/Log Pis Min -8.96722 +trainer/policy/mean Mean -0.0478219 +trainer/policy/mean Std 0.909506 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.79974 +trainer/policy/normal/std Std 0.677179 +trainer/policy/normal/std Max 5.38021 +trainer/policy/normal/std Min 0.237972 +trainer/policy/normal/log_std Mean 0.986468 +trainer/policy/normal/log_std Std 0.333888 +trainer/policy/normal/log_std Max 1.68273 +trainer/policy/normal/log_std Min -1.4356 +eval/num steps total 937446 +eval/num paths total 940 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0371843 +eval/Actions Std 0.913401 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66403 +time/logging (s) 0.00370054 +time/sampling batch (s) 0.776377 +time/saving (s) 0.003377 +time/training (s) 7.63297 +time/epoch (s) 11.0805 +time/total (s) 9740.99 +Epoch -61 +---------------------------------- --------------- +2022-05-10 15:53:23.922963 PDT | [1] Epoch -60 finished +---------------------------------- --------------- +epoch -60 +replay_buffer/size 999033 +trainer/num train calls 941000 +trainer/Policy Loss -19.737 +trainer/Log Pis Mean 24.3664 +trainer/Log Pis Std 13.2936 +trainer/Log Pis Max 67.9183 +trainer/Log Pis Min -8.59712 +trainer/policy/mean Mean -0.0460417 +trainer/policy/mean Std 0.902542 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.79584 +trainer/policy/normal/std Std 0.685524 +trainer/policy/normal/std Max 5.88471 +trainer/policy/normal/std Min 0.234119 +trainer/policy/normal/log_std Mean 0.982858 +trainer/policy/normal/log_std Std 0.343822 +trainer/policy/normal/log_std Max 1.77236 +trainer/policy/normal/log_std Min -1.45193 +eval/num steps total 938446 +eval/num paths total 941 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102039 +eval/Actions Std 0.911888 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59513 +time/logging (s) 0.00398158 +time/sampling batch (s) 0.275786 +time/saving (s) 0.0035528 +time/training (s) 8.29544 +time/epoch (s) 11.1739 +time/total (s) 9752.16 +Epoch -60 +---------------------------------- --------------- +2022-05-10 15:53:33.820547 PDT | [1] Epoch -59 finished +---------------------------------- --------------- +epoch -59 +replay_buffer/size 999033 +trainer/num train calls 942000 +trainer/Policy Loss -18.9084 +trainer/Log Pis Mean 24.4247 +trainer/Log Pis Std 13.2751 +trainer/Log Pis Max 71.603 +trainer/Log Pis Min -8.42056 +trainer/policy/mean Mean -0.0445844 +trainer/policy/mean Std 0.904702 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84927 +trainer/policy/normal/std Std 0.679449 +trainer/policy/normal/std Max 6.3134 +trainer/policy/normal/std Min 0.287276 +trainer/policy/normal/log_std Mean 1.0061 +trainer/policy/normal/log_std Std 0.323428 +trainer/policy/normal/log_std Max 1.84268 +trainer/policy/normal/log_std Min -1.24731 +eval/num steps total 939446 +eval/num paths total 942 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.180219 +eval/Actions Std 0.895764 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4993 +time/logging (s) 0.00378983 +time/sampling batch (s) 0.277337 +time/saving (s) 0.0034136 +time/training (s) 7.0882 +time/epoch (s) 9.87205 +time/total (s) 9762.04 +Epoch -59 +---------------------------------- --------------- +2022-05-10 15:53:44.699435 PDT | [1] Epoch -58 finished +---------------------------------- --------------- +epoch -58 +replay_buffer/size 999033 +trainer/num train calls 943000 +trainer/Policy Loss -18.0158 +trainer/Log Pis Mean 23.3952 +trainer/Log Pis Std 13.3994 +trainer/Log Pis Max 74.0155 +trainer/Log Pis Min -12.2611 +trainer/policy/mean Mean -0.0314011 +trainer/policy/mean Std 0.905113 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81475 +trainer/policy/normal/std Std 0.670924 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.291978 +trainer/policy/normal/log_std Mean 0.994312 +trainer/policy/normal/log_std Std 0.320813 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.23108 +eval/num steps total 940446 +eval/num paths total 943 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.157305 +eval/Actions Std 0.948197 +eval/Actions Max 0.999978 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64549 +time/logging (s) 0.00416237 +time/sampling batch (s) 0.785633 +time/saving (s) 0.00382164 +time/training (s) 7.41468 +time/epoch (s) 10.8538 +time/total (s) 9772.9 +Epoch -58 +---------------------------------- --------------- +2022-05-10 15:53:54.616236 PDT | [1] Epoch -57 finished +---------------------------------- ---------------- +epoch -57 +replay_buffer/size 999033 +trainer/num train calls 944000 +trainer/Policy Loss -19.9712 +trainer/Log Pis Mean 24.9683 +trainer/Log Pis Std 13.2784 +trainer/Log Pis Max 67.1475 +trainer/Log Pis Min -3.99803 +trainer/policy/mean Mean -0.0512511 +trainer/policy/mean Std 0.907744 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.77059 +trainer/policy/normal/std Std 0.657177 +trainer/policy/normal/std Max 6.26291 +trainer/policy/normal/std Min 0.277349 +trainer/policy/normal/log_std Mean 0.977259 +trainer/policy/normal/log_std Std 0.329061 +trainer/policy/normal/log_std Max 1.83465 +trainer/policy/normal/log_std Min -1.28248 +eval/num steps total 941446 +eval/num paths total 944 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 8.24184e-05 +eval/Actions Std 0.958578 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60306 +time/logging (s) 0.00396243 +time/sampling batch (s) 0.278719 +time/saving (s) 0.0037183 +time/training (s) 7.0016 +time/epoch (s) 9.89106 +time/total (s) 9782.79 +Epoch -57 +---------------------------------- ---------------- +2022-05-10 15:54:04.452236 PDT | [1] Epoch -56 finished +---------------------------------- --------------- +epoch -56 +replay_buffer/size 999033 +trainer/num train calls 945000 +trainer/Policy Loss -19.0663 +trainer/Log Pis Mean 24.2145 +trainer/Log Pis Std 13.7491 +trainer/Log Pis Max 68.1046 +trainer/Log Pis Min -8.58729 +trainer/policy/mean Mean -0.0362333 +trainer/policy/mean Std 0.90983 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.73468 +trainer/policy/normal/std Std 0.645376 +trainer/policy/normal/std Max 5.43402 +trainer/policy/normal/std Min 0.298775 +trainer/policy/normal/log_std Mean 0.964868 +trainer/policy/normal/log_std Std 0.325576 +trainer/policy/normal/log_std Max 1.69268 +trainer/policy/normal/log_std Min -1.20807 +eval/num steps total 942446 +eval/num paths total 945 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0494889 +eval/Actions Std 0.908103 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59142 +time/logging (s) 0.00379853 +time/sampling batch (s) 0.282603 +time/saving (s) 0.00347491 +time/training (s) 6.92902 +time/epoch (s) 9.81032 +time/total (s) 9792.6 +Epoch -56 +---------------------------------- --------------- +2022-05-10 15:54:14.229019 PDT | [1] Epoch -55 finished +---------------------------------- --------------- +epoch -55 +replay_buffer/size 999033 +trainer/num train calls 946000 +trainer/Policy Loss -18.6172 +trainer/Log Pis Mean 24.4211 +trainer/Log Pis Std 13.8125 +trainer/Log Pis Max 74.3609 +trainer/Log Pis Min -8.78768 +trainer/policy/mean Mean -0.0420289 +trainer/policy/mean Std 0.906089 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.782 +trainer/policy/normal/std Std 0.657693 +trainer/policy/normal/std Max 6.03715 +trainer/policy/normal/std Min 0.223686 +trainer/policy/normal/log_std Mean 0.981649 +trainer/policy/normal/log_std Std 0.329751 +trainer/policy/normal/log_std Max 1.79793 +trainer/policy/normal/log_std Min -1.49751 +eval/num steps total 943446 +eval/num paths total 946 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.278454 +eval/Actions Std 0.860873 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54783 +time/logging (s) 0.00374262 +time/sampling batch (s) 0.275435 +time/saving (s) 0.00344756 +time/training (s) 6.92133 +time/epoch (s) 9.75179 +time/total (s) 9802.36 +Epoch -55 +---------------------------------- --------------- +2022-05-10 15:54:24.423317 PDT | [1] Epoch -54 finished +---------------------------------- --------------- +epoch -54 +replay_buffer/size 999033 +trainer/num train calls 947000 +trainer/Policy Loss -18.6484 +trainer/Log Pis Mean 25.6924 +trainer/Log Pis Std 13.4442 +trainer/Log Pis Max 65.8569 +trainer/Log Pis Min -6.96309 +trainer/policy/mean Mean -0.0474057 +trainer/policy/mean Std 0.90534 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.84485 +trainer/policy/normal/std Std 0.711026 +trainer/policy/normal/std Max 6.34867 +trainer/policy/normal/std Min 0.268993 +trainer/policy/normal/log_std Mean 0.998845 +trainer/policy/normal/log_std Std 0.349217 +trainer/policy/normal/log_std Max 1.84825 +trainer/policy/normal/log_std Min -1.31307 +eval/num steps total 944446 +eval/num paths total 947 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0406729 +eval/Actions Std 0.919087 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6572 +time/logging (s) 0.00375896 +time/sampling batch (s) 0.279995 +time/saving (s) 0.00345433 +time/training (s) 7.22469 +time/epoch (s) 10.1691 +time/total (s) 9812.53 +Epoch -54 +---------------------------------- --------------- +2022-05-10 15:54:34.995275 PDT | [1] Epoch -53 finished +---------------------------------- --------------- +epoch -53 +replay_buffer/size 999033 +trainer/num train calls 948000 +trainer/Policy Loss -19.2813 +trainer/Log Pis Mean 23.9047 +trainer/Log Pis Std 14.2958 +trainer/Log Pis Max 77.3512 +trainer/Log Pis Min -9.97184 +trainer/policy/mean Mean -0.0575247 +trainer/policy/mean Std 0.90268 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81203 +trainer/policy/normal/std Std 0.667639 +trainer/policy/normal/std Max 5.32373 +trainer/policy/normal/std Min 0.245411 +trainer/policy/normal/log_std Mean 0.993073 +trainer/policy/normal/log_std Std 0.322965 +trainer/policy/normal/log_std Max 1.67217 +trainer/policy/normal/log_std Min -1.40482 +eval/num steps total 945446 +eval/num paths total 948 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119193 +eval/Actions Std 0.916541 +eval/Actions Max 0.999993 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61197 +time/logging (s) 0.00385943 +time/sampling batch (s) 0.527654 +time/saving (s) 0.00341608 +time/training (s) 7.3999 +time/epoch (s) 10.5468 +time/total (s) 9823.08 +Epoch -53 +---------------------------------- --------------- +2022-05-10 15:54:44.666219 PDT | [1] Epoch -52 finished +---------------------------------- --------------- +epoch -52 +replay_buffer/size 999033 +trainer/num train calls 949000 +trainer/Policy Loss -19.5509 +trainer/Log Pis Mean 24.2175 +trainer/Log Pis Std 14.3456 +trainer/Log Pis Max 75.9528 +trainer/Log Pis Min -9.16426 +trainer/policy/mean Mean -0.0555968 +trainer/policy/mean Std 0.91258 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.80502 +trainer/policy/normal/std Std 0.64825 +trainer/policy/normal/std Max 6.9876 +trainer/policy/normal/std Min 0.296172 +trainer/policy/normal/log_std Mean 0.993803 +trainer/policy/normal/log_std Std 0.308462 +trainer/policy/normal/log_std Max 1.94414 +trainer/policy/normal/log_std Min -1.21682 +eval/num steps total 946446 +eval/num paths total 949 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.244598 +eval/Actions Std 0.84796 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70869 +time/logging (s) 0.00405986 +time/sampling batch (s) 0.276099 +time/saving (s) 0.00378942 +time/training (s) 6.65345 +time/epoch (s) 9.64609 +time/total (s) 9832.73 +Epoch -52 +---------------------------------- --------------- +2022-05-10 15:54:55.578859 PDT | [1] Epoch -51 finished +---------------------------------- --------------- +epoch -51 +replay_buffer/size 999033 +trainer/num train calls 950000 +trainer/Policy Loss -19.7004 +trainer/Log Pis Mean 25.1865 +trainer/Log Pis Std 13.7753 +trainer/Log Pis Max 73.3216 +trainer/Log Pis Min -10.3759 +trainer/policy/mean Mean -0.043007 +trainer/policy/mean Std 0.908519 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.86627 +trainer/policy/normal/std Std 0.70624 +trainer/policy/normal/std Max 5.86046 +trainer/policy/normal/std Min 0.24267 +trainer/policy/normal/log_std Mean 1.00737 +trainer/policy/normal/log_std Std 0.345739 +trainer/policy/normal/log_std Max 1.76823 +trainer/policy/normal/log_std Min -1.41605 +eval/num steps total 947446 +eval/num paths total 950 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.148507 +eval/Actions Std 0.866828 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49044 +time/logging (s) 0.0037783 +time/sampling batch (s) 0.277303 +time/saving (s) 0.00357873 +time/training (s) 8.11156 +time/epoch (s) 10.8867 +time/total (s) 9843.62 +Epoch -51 +---------------------------------- --------------- +2022-05-10 15:55:05.135237 PDT | [1] Epoch -50 finished +---------------------------------- --------------- +epoch -50 +replay_buffer/size 999033 +trainer/num train calls 951000 +trainer/Policy Loss -19.0888 +trainer/Log Pis Mean 24.5173 +trainer/Log Pis Std 13.0103 +trainer/Log Pis Max 62.6599 +trainer/Log Pis Min -4.36429 +trainer/policy/mean Mean -0.0237384 +trainer/policy/mean Std 0.905356 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.80483 +trainer/policy/normal/std Std 0.670913 +trainer/policy/normal/std Max 5.83936 +trainer/policy/normal/std Min 0.250743 +trainer/policy/normal/log_std Mean 0.989697 +trainer/policy/normal/log_std Std 0.327412 +trainer/policy/normal/log_std Max 1.76462 +trainer/policy/normal/log_std Min -1.38333 +eval/num steps total 948446 +eval/num paths total 951 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.30189 +eval/Actions Std 0.839776 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77585 +time/logging (s) 0.00369947 +time/sampling batch (s) 0.276533 +time/saving (s) 0.00348163 +time/training (s) 6.4714 +time/epoch (s) 9.53096 +time/total (s) 9853.15 +Epoch -50 +---------------------------------- --------------- +2022-05-10 15:55:14.310287 PDT | [1] Epoch -49 finished +---------------------------------- --------------- +epoch -49 +replay_buffer/size 999033 +trainer/num train calls 952000 +trainer/Policy Loss -20.3592 +trainer/Log Pis Mean 24.8961 +trainer/Log Pis Std 13.1139 +trainer/Log Pis Max 71.8465 +trainer/Log Pis Min -5.95446 +trainer/policy/mean Mean -0.0559934 +trainer/policy/mean Std 0.90551 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7771 +trainer/policy/normal/std Std 0.672528 +trainer/policy/normal/std Max 5.90274 +trainer/policy/normal/std Min 0.254442 +trainer/policy/normal/log_std Mean 0.979034 +trainer/policy/normal/log_std Std 0.329139 +trainer/policy/normal/log_std Max 1.77542 +trainer/policy/normal/log_std Min -1.36868 +eval/num steps total 949446 +eval/num paths total 952 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0495784 +eval/Actions Std 0.904068 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64859 +time/logging (s) 0.00374298 +time/sampling batch (s) 0.275062 +time/saving (s) 0.00347089 +time/training (s) 6.21913 +time/epoch (s) 9.15 +time/total (s) 9862.31 +Epoch -49 +---------------------------------- --------------- +2022-05-10 15:55:25.013277 PDT | [1] Epoch -48 finished +---------------------------------- --------------- +epoch -48 +replay_buffer/size 999033 +trainer/num train calls 953000 +trainer/Policy Loss -19.5744 +trainer/Log Pis Mean 23.8334 +trainer/Log Pis Std 13.299 +trainer/Log Pis Max 66.6869 +trainer/Log Pis Min -7.04089 +trainer/policy/mean Mean -0.0291593 +trainer/policy/mean Std 0.906374 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.72479 +trainer/policy/normal/std Std 0.704619 +trainer/policy/normal/std Max 6.34705 +trainer/policy/normal/std Min 0.203389 +trainer/policy/normal/log_std Mean 0.951431 +trainer/policy/normal/log_std Std 0.367429 +trainer/policy/normal/log_std Max 1.84799 +trainer/policy/normal/log_std Min -1.59264 +eval/num steps total 950446 +eval/num paths total 953 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121595 +eval/Actions Std 0.871775 +eval/Actions Max 0.999968 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78663 +time/logging (s) 0.00403662 +time/sampling batch (s) 0.534306 +time/saving (s) 0.00347002 +time/training (s) 7.34925 +time/epoch (s) 10.6777 +time/total (s) 9872.99 +Epoch -48 +---------------------------------- --------------- +2022-05-10 15:55:34.962595 PDT | [1] Epoch -47 finished +---------------------------------- --------------- +epoch -47 +replay_buffer/size 999033 +trainer/num train calls 954000 +trainer/Policy Loss -19.1796 +trainer/Log Pis Mean 24.6415 +trainer/Log Pis Std 13.8678 +trainer/Log Pis Max 71.9178 +trainer/Log Pis Min -12.3964 +trainer/policy/mean Mean -0.0289063 +trainer/policy/mean Std 0.904965 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.78567 +trainer/policy/normal/std Std 0.692224 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.290471 +trainer/policy/normal/log_std Mean 0.979574 +trainer/policy/normal/log_std Std 0.339861 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.23625 +eval/num steps total 951446 +eval/num paths total 954 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0303122 +eval/Actions Std 0.907617 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63996 +time/logging (s) 0.00427411 +time/sampling batch (s) 0.281936 +time/saving (s) 0.00400296 +time/training (s) 6.99333 +time/epoch (s) 9.9235 +time/total (s) 9882.91 +Epoch -47 +---------------------------------- --------------- +2022-05-10 15:55:45.145582 PDT | [1] Epoch -46 finished +---------------------------------- --------------- +epoch -46 +replay_buffer/size 999033 +trainer/num train calls 955000 +trainer/Policy Loss -19.1975 +trainer/Log Pis Mean 25.5046 +trainer/Log Pis Std 13.3298 +trainer/Log Pis Max 88.1924 +trainer/Log Pis Min -3.51986 +trainer/policy/mean Mean -0.0222033 +trainer/policy/mean Std 0.909147 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.81709 +trainer/policy/normal/std Std 0.684561 +trainer/policy/normal/std Max 5.95048 +trainer/policy/normal/std Min 0.281524 +trainer/policy/normal/log_std Mean 0.990832 +trainer/policy/normal/log_std Std 0.343866 +trainer/policy/normal/log_std Max 1.78347 +trainer/policy/normal/log_std Min -1.26754 +eval/num steps total 952446 +eval/num paths total 955 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0903773 +eval/Actions Std 0.920652 +eval/Actions Max 0.99999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45755 +time/logging (s) 0.00385765 +time/sampling batch (s) 0.28154 +time/saving (s) 0.00360891 +time/training (s) 7.41018 +time/epoch (s) 10.1567 +time/total (s) 9893.07 +Epoch -46 +---------------------------------- --------------- +2022-05-10 15:55:54.957854 PDT | [1] Epoch -45 finished +---------------------------------- --------------- +epoch -45 +replay_buffer/size 999033 +trainer/num train calls 956000 +trainer/Policy Loss -18.4933 +trainer/Log Pis Mean 24.026 +trainer/Log Pis Std 13.0735 +trainer/Log Pis Max 62.5174 +trainer/Log Pis Min -9.25025 +trainer/policy/mean Mean -0.0273351 +trainer/policy/mean Std 0.903794 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.77461 +trainer/policy/normal/std Std 0.672443 +trainer/policy/normal/std Max 6.3137 +trainer/policy/normal/std Min 0.257683 +trainer/policy/normal/log_std Mean 0.977069 +trainer/policy/normal/log_std Std 0.337082 +trainer/policy/normal/log_std Max 1.84272 +trainer/policy/normal/log_std Min -1.35603 +eval/num steps total 953446 +eval/num paths total 956 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.226733 +eval/Actions Std 0.902646 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64722 +time/logging (s) 0.00378853 +time/sampling batch (s) 0.281642 +time/saving (s) 0.00346557 +time/training (s) 6.85051 +time/epoch (s) 9.78663 +time/total (s) 9902.86 +Epoch -45 +---------------------------------- --------------- +2022-05-10 15:56:04.606483 PDT | [1] Epoch -44 finished +---------------------------------- --------------- +epoch -44 +replay_buffer/size 999033 +trainer/num train calls 957000 +trainer/Policy Loss -21.4821 +trainer/Log Pis Mean 24.0739 +trainer/Log Pis Std 12.989 +trainer/Log Pis Max 66.1587 +trainer/Log Pis Min -8.56025 +trainer/policy/mean Mean -0.0300231 +trainer/policy/mean Std 0.908918 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.85309 +trainer/policy/normal/std Std 0.678598 +trainer/policy/normal/std Max 5.47572 +trainer/policy/normal/std Min 0.252208 +trainer/policy/normal/log_std Mean 1.0073 +trainer/policy/normal/log_std Std 0.324474 +trainer/policy/normal/log_std Max 1.70032 +trainer/policy/normal/log_std Min -1.3775 +eval/num steps total 954446 +eval/num paths total 957 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00525653 +eval/Actions Std 0.96709 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56762 +time/logging (s) 0.00383474 +time/sampling batch (s) 0.280857 +time/saving (s) 0.00347393 +time/training (s) 6.76745 +time/epoch (s) 9.62324 +time/total (s) 9912.49 +Epoch -44 +---------------------------------- --------------- +2022-05-10 15:56:15.869297 PDT | [1] Epoch -43 finished +---------------------------------- --------------- +epoch -43 +replay_buffer/size 999033 +trainer/num train calls 958000 +trainer/Policy Loss -19.867 +trainer/Log Pis Mean 24.1009 +trainer/Log Pis Std 13.4352 +trainer/Log Pis Max 86.0139 +trainer/Log Pis Min -7.60633 +trainer/policy/mean Mean -0.0377304 +trainer/policy/mean Std 0.904804 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.8014 +trainer/policy/normal/std Std 0.713774 +trainer/policy/normal/std Max 6.32923 +trainer/policy/normal/std Min 0.300353 +trainer/policy/normal/log_std Mean 0.98205 +trainer/policy/normal/log_std Std 0.353683 +trainer/policy/normal/log_std Max 1.84518 +trainer/policy/normal/log_std Min -1.2028 +eval/num steps total 955446 +eval/num paths total 958 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103177 +eval/Actions Std 0.913904 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6427 +time/logging (s) 0.00434516 +time/sampling batch (s) 1.02895 +time/saving (s) 0.00406476 +time/training (s) 7.55776 +time/epoch (s) 11.2378 +time/total (s) 9923.73 +Epoch -43 +---------------------------------- --------------- +2022-05-10 15:56:25.635440 PDT | [1] Epoch -42 finished +---------------------------------- --------------- +epoch -42 +replay_buffer/size 999033 +trainer/num train calls 959000 +trainer/Policy Loss -19.9578 +trainer/Log Pis Mean 23.4864 +trainer/Log Pis Std 12.337 +trainer/Log Pis Max 60.7095 +trainer/Log Pis Min -6.87992 +trainer/policy/mean Mean -0.0271658 +trainer/policy/mean Std 0.905681 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76475 +trainer/policy/normal/std Std 0.655898 +trainer/policy/normal/std Max 5.49939 +trainer/policy/normal/std Min 0.243518 +trainer/policy/normal/log_std Mean 0.97623 +trainer/policy/normal/log_std Std 0.322478 +trainer/policy/normal/log_std Max 1.70464 +trainer/policy/normal/log_std Min -1.41256 +eval/num steps total 956446 +eval/num paths total 959 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.305503 +eval/Actions Std 0.875497 +eval/Actions Max 0.99998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68554 +time/logging (s) 0.00390596 +time/sampling batch (s) 0.526264 +time/saving (s) 0.00347753 +time/training (s) 6.52098 +time/epoch (s) 9.74016 +time/total (s) 9933.47 +Epoch -42 +---------------------------------- --------------- +2022-05-10 15:56:34.822526 PDT | [1] Epoch -41 finished +---------------------------------- --------------- +epoch -41 +replay_buffer/size 999033 +trainer/num train calls 960000 +trainer/Policy Loss -17.6627 +trainer/Log Pis Mean 23.7263 +trainer/Log Pis Std 13.6828 +trainer/Log Pis Max 70.8963 +trainer/Log Pis Min -6.67706 +trainer/policy/mean Mean -0.0233647 +trainer/policy/mean Std 0.904965 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.80629 +trainer/policy/normal/std Std 0.670639 +trainer/policy/normal/std Max 6.49889 +trainer/policy/normal/std Min 0.239935 +trainer/policy/normal/log_std Mean 0.991078 +trainer/policy/normal/log_std Std 0.323188 +trainer/policy/normal/log_std Max 1.87163 +trainer/policy/normal/log_std Min -1.42739 +eval/num steps total 957446 +eval/num paths total 960 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.439459 +eval/Actions Std 0.692511 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51576 +time/logging (s) 0.00426948 +time/sampling batch (s) 0.528814 +time/saving (s) 0.0041531 +time/training (s) 6.10918 +time/epoch (s) 9.16217 +time/total (s) 9942.64 +Epoch -41 +---------------------------------- --------------- +2022-05-10 15:56:45.590315 PDT | [1] Epoch -40 finished +---------------------------------- --------------- +epoch -40 +replay_buffer/size 999033 +trainer/num train calls 961000 +trainer/Policy Loss -19.7102 +trainer/Log Pis Mean 23.5656 +trainer/Log Pis Std 13.4937 +trainer/Log Pis Max 71.2091 +trainer/Log Pis Min -5.34704 +trainer/policy/mean Mean -0.0199411 +trainer/policy/mean Std 0.90405 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77931 +trainer/policy/normal/std Std 0.690824 +trainer/policy/normal/std Max 5.40276 +trainer/policy/normal/std Min 0.200157 +trainer/policy/normal/log_std Mean 0.975208 +trainer/policy/normal/log_std Std 0.352704 +trainer/policy/normal/log_std Max 1.68691 +trainer/policy/normal/log_std Min -1.60865 +eval/num steps total 958446 +eval/num paths total 961 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.341946 +eval/Actions Std 0.828119 +eval/Actions Max 0.999978 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75816 +time/logging (s) 0.00424017 +time/sampling batch (s) 0.779464 +time/saving (s) 0.00357737 +time/training (s) 7.19638 +time/epoch (s) 10.7418 +time/total (s) 9953.38 +Epoch -40 +---------------------------------- --------------- +2022-05-10 15:56:55.544079 PDT | [1] Epoch -39 finished +---------------------------------- --------------- +epoch -39 +replay_buffer/size 999033 +trainer/num train calls 962000 +trainer/Policy Loss -20.8819 +trainer/Log Pis Mean 24.3531 +trainer/Log Pis Std 13.4298 +trainer/Log Pis Max 65.1034 +trainer/Log Pis Min -8.24126 +trainer/policy/mean Mean -0.0302406 +trainer/policy/mean Std 0.903526 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79224 +trainer/policy/normal/std Std 0.691473 +trainer/policy/normal/std Max 5.47473 +trainer/policy/normal/std Min 0.280723 +trainer/policy/normal/log_std Mean 0.978299 +trainer/policy/normal/log_std Std 0.360961 +trainer/policy/normal/log_std Max 1.70014 +trainer/policy/normal/log_std Min -1.27039 +eval/num steps total 959446 +eval/num paths total 962 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116762 +eval/Actions Std 0.901693 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65641 +time/logging (s) 0.00388937 +time/sampling batch (s) 0.280238 +time/saving (s) 0.00360003 +time/training (s) 6.98318 +time/epoch (s) 9.92732 +time/total (s) 9963.31 +Epoch -39 +---------------------------------- --------------- +2022-05-10 15:57:06.370571 PDT | [1] Epoch -38 finished +---------------------------------- --------------- +epoch -38 +replay_buffer/size 999033 +trainer/num train calls 963000 +trainer/Policy Loss -19.4554 +trainer/Log Pis Mean 24.0469 +trainer/Log Pis Std 12.7854 +trainer/Log Pis Max 65.088 +trainer/Log Pis Min -6.77567 +trainer/policy/mean Mean -0.0452405 +trainer/policy/mean Std 0.906444 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.80021 +trainer/policy/normal/std Std 0.665812 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.212868 +trainer/policy/normal/log_std Mean 0.989358 +trainer/policy/normal/log_std Std 0.32046 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.54708 +eval/num steps total 960446 +eval/num paths total 963 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.145154 +eval/Actions Std 0.918243 +eval/Actions Max 0.999996 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62452 +time/logging (s) 0.00431254 +time/sampling batch (s) 0.530378 +time/saving (s) 0.0041231 +time/training (s) 7.63805 +time/epoch (s) 10.8014 +time/total (s) 9974.12 +Epoch -38 +---------------------------------- --------------- +2022-05-10 15:57:16.338418 PDT | [1] Epoch -37 finished +---------------------------------- --------------- +epoch -37 +replay_buffer/size 999033 +trainer/num train calls 964000 +trainer/Policy Loss -18.1192 +trainer/Log Pis Mean 24.9294 +trainer/Log Pis Std 13.6281 +trainer/Log Pis Max 74.2476 +trainer/Log Pis Min -7.79769 +trainer/policy/mean Mean -0.04989 +trainer/policy/mean Std 0.904889 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84152 +trainer/policy/normal/std Std 0.703339 +trainer/policy/normal/std Max 6.88617 +trainer/policy/normal/std Min 0.315859 +trainer/policy/normal/log_std Mean 1.00012 +trainer/policy/normal/log_std Std 0.336578 +trainer/policy/normal/log_std Max 1.92951 +trainer/policy/normal/log_std Min -1.15246 +eval/num steps total 961446 +eval/num paths total 964 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0075567 +eval/Actions Std 0.906431 +eval/Actions Max 0.999988 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66341 +time/logging (s) 0.00379849 +time/sampling batch (s) 0.279579 +time/saving (s) 0.00349809 +time/training (s) 6.99114 +time/epoch (s) 9.94143 +time/total (s) 9984.06 +Epoch -37 +---------------------------------- --------------- +2022-05-10 15:57:26.637854 PDT | [1] Epoch -36 finished +---------------------------------- --------------- +epoch -36 +replay_buffer/size 999033 +trainer/num train calls 965000 +trainer/Policy Loss -19.0589 +trainer/Log Pis Mean 25.0382 +trainer/Log Pis Std 13.4555 +trainer/Log Pis Max 76.3764 +trainer/Log Pis Min -9.11414 +trainer/policy/mean Mean -0.0370871 +trainer/policy/mean Std 0.907395 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7636 +trainer/policy/normal/std Std 0.690107 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.265853 +trainer/policy/normal/log_std Mean 0.972053 +trainer/policy/normal/log_std Std 0.336924 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.32481 +eval/num steps total 962446 +eval/num paths total 965 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00069768 +eval/Actions Std 0.853829 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68232 +time/logging (s) 0.00386693 +time/sampling batch (s) 0.526334 +time/saving (s) 0.00339845 +time/training (s) 7.05833 +time/epoch (s) 10.2743 +time/total (s) 9994.34 +Epoch -36 +---------------------------------- --------------- +2022-05-10 15:57:37.973889 PDT | [1] Epoch -35 finished +---------------------------------- --------------- +epoch -35 +replay_buffer/size 999033 +trainer/num train calls 966000 +trainer/Policy Loss -19.5783 +trainer/Log Pis Mean 23.2104 +trainer/Log Pis Std 12.8 +trainer/Log Pis Max 59.5222 +trainer/Log Pis Min -12.7622 +trainer/policy/mean Mean -0.0464363 +trainer/policy/mean Std 0.907326 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79839 +trainer/policy/normal/std Std 0.661871 +trainer/policy/normal/std Max 5.46595 +trainer/policy/normal/std Min 0.295268 +trainer/policy/normal/log_std Mean 0.987784 +trainer/policy/normal/log_std Std 0.326161 +trainer/policy/normal/log_std Max 1.69854 +trainer/policy/normal/log_std Min -1.21987 +eval/num steps total 963446 +eval/num paths total 966 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.446489 +eval/Actions Std 0.813166 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66023 +time/logging (s) 0.00425985 +time/sampling batch (s) 1.04073 +time/saving (s) 0.00413193 +time/training (s) 7.60107 +time/epoch (s) 11.3104 +time/total (s) 10005.7 +Epoch -35 +---------------------------------- --------------- +2022-05-10 15:57:49.226761 PDT | [1] Epoch -34 finished +---------------------------------- --------------- +epoch -34 +replay_buffer/size 999033 +trainer/num train calls 967000 +trainer/Policy Loss -19.9586 +trainer/Log Pis Mean 24.9874 +trainer/Log Pis Std 13.5935 +trainer/Log Pis Max 67.5136 +trainer/Log Pis Min -12.1587 +trainer/policy/mean Mean -0.046616 +trainer/policy/mean Std 0.909772 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79921 +trainer/policy/normal/std Std 0.65789 +trainer/policy/normal/std Max 6.1807 +trainer/policy/normal/std Min 0.232637 +trainer/policy/normal/log_std Mean 0.989549 +trainer/policy/normal/log_std Std 0.319376 +trainer/policy/normal/log_std Max 1.82143 +trainer/policy/normal/log_std Min -1.45828 +eval/num steps total 964446 +eval/num paths total 967 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00719621 +eval/Actions Std 0.908558 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57339 +time/logging (s) 0.00391181 +time/sampling batch (s) 0.286273 +time/saving (s) 0.00363426 +time/training (s) 8.35883 +time/epoch (s) 11.226 +time/total (s) 10016.9 +Epoch -34 +---------------------------------- --------------- +2022-05-10 15:58:00.300456 PDT | [1] Epoch -33 finished +---------------------------------- --------------- +epoch -33 +replay_buffer/size 999033 +trainer/num train calls 968000 +trainer/Policy Loss -19.7105 +trainer/Log Pis Mean 25.2731 +trainer/Log Pis Std 13.6641 +trainer/Log Pis Max 76.44 +trainer/Log Pis Min -5.56217 +trainer/policy/mean Mean -0.0214058 +trainer/policy/mean Std 0.909861 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.79241 +trainer/policy/normal/std Std 0.697986 +trainer/policy/normal/std Max 6.53415 +trainer/policy/normal/std Min 0.315728 +trainer/policy/normal/log_std Mean 0.98131 +trainer/policy/normal/log_std Std 0.34217 +trainer/policy/normal/log_std Max 1.87704 +trainer/policy/normal/log_std Min -1.15288 +eval/num steps total 965446 +eval/num paths total 968 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.166369 +eval/Actions Std 0.952495 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67159 +time/logging (s) 0.00395258 +time/sampling batch (s) 0.28881 +time/saving (s) 0.00354303 +time/training (s) 8.07975 +time/epoch (s) 11.0477 +time/total (s) 10027.9 +Epoch -33 +---------------------------------- --------------- +2022-05-10 15:58:10.692185 PDT | [1] Epoch -32 finished +---------------------------------- --------------- +epoch -32 +replay_buffer/size 999033 +trainer/num train calls 969000 +trainer/Policy Loss -19.9158 +trainer/Log Pis Mean 24.0559 +trainer/Log Pis Std 13.3613 +trainer/Log Pis Max 81.706 +trainer/Log Pis Min -8.21452 +trainer/policy/mean Mean -0.0382181 +trainer/policy/mean Std 0.908088 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85312 +trainer/policy/normal/std Std 0.701441 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.244101 +trainer/policy/normal/log_std Mean 1.00495 +trainer/policy/normal/log_std Std 0.333702 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.41017 +eval/num steps total 966446 +eval/num paths total 969 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.17817 +eval/Actions Std 0.892056 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7255 +time/logging (s) 0.00387158 +time/sampling batch (s) 0.528124 +time/saving (s) 0.00344961 +time/training (s) 7.10534 +time/epoch (s) 10.3663 +time/total (s) 10038.3 +Epoch -32 +---------------------------------- --------------- +2022-05-10 15:58:20.973448 PDT | [1] Epoch -31 finished +---------------------------------- --------------- +epoch -31 +replay_buffer/size 999033 +trainer/num train calls 970000 +trainer/Policy Loss -18.9598 +trainer/Log Pis Mean 23.6112 +trainer/Log Pis Std 13.2978 +trainer/Log Pis Max 61.3492 +trainer/Log Pis Min -8.11906 +trainer/policy/mean Mean -0.0608087 +trainer/policy/mean Std 0.903136 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.76843 +trainer/policy/normal/std Std 0.702662 +trainer/policy/normal/std Max 5.56639 +trainer/policy/normal/std Min 0.221092 +trainer/policy/normal/log_std Mean 0.968746 +trainer/policy/normal/log_std Std 0.362447 +trainer/policy/normal/log_std Max 1.71675 +trainer/policy/normal/log_std Min -1.50918 +eval/num steps total 967446 +eval/num paths total 970 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114546 +eval/Actions Std 0.919442 +eval/Actions Max 0.99998 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55226 +time/logging (s) 0.00433624 +time/sampling batch (s) 0.279841 +time/saving (s) 0.00413367 +time/training (s) 7.41552 +time/epoch (s) 10.2561 +time/total (s) 10048.6 +Epoch -31 +---------------------------------- --------------- +2022-05-10 15:58:31.977911 PDT | [1] Epoch -30 finished +---------------------------------- --------------- +epoch -30 +replay_buffer/size 999033 +trainer/num train calls 971000 +trainer/Policy Loss -19.3921 +trainer/Log Pis Mean 24.2169 +trainer/Log Pis Std 13.7336 +trainer/Log Pis Max 66.2462 +trainer/Log Pis Min -6.4573 +trainer/policy/mean Mean -0.0222402 +trainer/policy/mean Std 0.911352 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77836 +trainer/policy/normal/std Std 0.663423 +trainer/policy/normal/std Max 5.38963 +trainer/policy/normal/std Min 0.291762 +trainer/policy/normal/log_std Mean 0.978204 +trainer/policy/normal/log_std Std 0.340603 +trainer/policy/normal/log_std Max 1.68448 +trainer/policy/normal/log_std Min -1.23182 +eval/num steps total 968446 +eval/num paths total 971 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.464377 +eval/Actions Std 0.802675 +eval/Actions Max 0.999993 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64914 +time/logging (s) 0.004396 +time/sampling batch (s) 0.28823 +time/saving (s) 0.00345044 +time/training (s) 8.0326 +time/epoch (s) 10.9778 +time/total (s) 10059.5 +Epoch -30 +---------------------------------- --------------- +2022-05-10 15:58:42.156299 PDT | [1] Epoch -29 finished +---------------------------------- --------------- +epoch -29 +replay_buffer/size 999033 +trainer/num train calls 972000 +trainer/Policy Loss -19.4046 +trainer/Log Pis Mean 25.119 +trainer/Log Pis Std 13.1885 +trainer/Log Pis Max 78.9169 +trainer/Log Pis Min -3.21387 +trainer/policy/mean Mean -0.0427852 +trainer/policy/mean Std 0.911116 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83514 +trainer/policy/normal/std Std 0.675141 +trainer/policy/normal/std Max 5.41054 +trainer/policy/normal/std Min 0.270996 +trainer/policy/normal/log_std Mean 1.00172 +trainer/policy/normal/log_std Std 0.320155 +trainer/policy/normal/log_std Max 1.68835 +trainer/policy/normal/log_std Min -1.30565 +eval/num steps total 969446 +eval/num paths total 972 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196377 +eval/Actions Std 0.894648 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69331 +time/logging (s) 0.00416375 +time/sampling batch (s) 0.275804 +time/saving (s) 0.00413172 +time/training (s) 7.17465 +time/epoch (s) 10.1521 +time/total (s) 10069.7 +Epoch -29 +---------------------------------- --------------- +2022-05-10 15:58:52.737659 PDT | [1] Epoch -28 finished +---------------------------------- --------------- +epoch -28 +replay_buffer/size 999033 +trainer/num train calls 973000 +trainer/Policy Loss -21.0183 +trainer/Log Pis Mean 24.863 +trainer/Log Pis Std 13.4936 +trainer/Log Pis Max 77.9061 +trainer/Log Pis Min -5.36404 +trainer/policy/mean Mean -0.0219722 +trainer/policy/mean Std 0.910566 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81023 +trainer/policy/normal/std Std 0.668502 +trainer/policy/normal/std Max 5.7556 +trainer/policy/normal/std Min 0.267858 +trainer/policy/normal/log_std Mean 0.991118 +trainer/policy/normal/log_std Std 0.331181 +trainer/policy/normal/log_std Max 1.75017 +trainer/policy/normal/log_std Min -1.3173 +eval/num steps total 970446 +eval/num paths total 973 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.359763 +eval/Actions Std 0.848048 +eval/Actions Max 0.999995 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57889 +time/logging (s) 0.00402504 +time/sampling batch (s) 0.2761 +time/saving (s) 0.00397817 +time/training (s) 7.69244 +time/epoch (s) 10.5554 +time/total (s) 10080.3 +Epoch -28 +---------------------------------- --------------- +2022-05-10 15:59:02.754021 PDT | [1] Epoch -27 finished +---------------------------------- --------------- +epoch -27 +replay_buffer/size 999033 +trainer/num train calls 974000 +trainer/Policy Loss -19.6758 +trainer/Log Pis Mean 24.525 +trainer/Log Pis Std 13.273 +trainer/Log Pis Max 69.4426 +trainer/Log Pis Min -5.1365 +trainer/policy/mean Mean -0.0300812 +trainer/policy/mean Std 0.907881 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82053 +trainer/policy/normal/std Std 0.67488 +trainer/policy/normal/std Max 6.62723 +trainer/policy/normal/std Min 0.321176 +trainer/policy/normal/log_std Mean 0.995956 +trainer/policy/normal/log_std Std 0.322695 +trainer/policy/normal/log_std Max 1.89119 +trainer/policy/normal/log_std Min -1.13576 +eval/num steps total 971446 +eval/num paths total 974 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249832 +eval/Actions Std 0.892543 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49198 +time/logging (s) 0.00427423 +time/sampling batch (s) 0.281615 +time/saving (s) 0.00397065 +time/training (s) 7.20857 +time/epoch (s) 9.99042 +time/total (s) 10090.3 +Epoch -27 +---------------------------------- --------------- +2022-05-10 15:59:12.504451 PDT | [1] Epoch -26 finished +---------------------------------- --------------- +epoch -26 +replay_buffer/size 999033 +trainer/num train calls 975000 +trainer/Policy Loss -19.4038 +trainer/Log Pis Mean 24.1095 +trainer/Log Pis Std 13.5655 +trainer/Log Pis Max 67.5914 +trainer/Log Pis Min -8.96409 +trainer/policy/mean Mean -0.0433513 +trainer/policy/mean Std 0.906186 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.80217 +trainer/policy/normal/std Std 0.704112 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.317835 +trainer/policy/normal/log_std Mean 0.9841 +trainer/policy/normal/log_std Std 0.345474 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.14622 +eval/num steps total 972446 +eval/num paths total 975 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0850579 +eval/Actions Std 0.88064 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52792 +time/logging (s) 0.00407773 +time/sampling batch (s) 0.283197 +time/saving (s) 0.00429833 +time/training (s) 6.90407 +time/epoch (s) 9.72356 +time/total (s) 10100 +Epoch -26 +---------------------------------- --------------- +2022-05-10 15:59:23.586908 PDT | [1] Epoch -25 finished +---------------------------------- --------------- +epoch -25 +replay_buffer/size 999033 +trainer/num train calls 976000 +trainer/Policy Loss -19.3998 +trainer/Log Pis Mean 25.2816 +trainer/Log Pis Std 14.2772 +trainer/Log Pis Max 80.1611 +trainer/Log Pis Min -11.4076 +trainer/policy/mean Mean -0.0550981 +trainer/policy/mean Std 0.904925 +trainer/policy/mean Max 1 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81804 +trainer/policy/normal/std Std 0.677036 +trainer/policy/normal/std Max 7.07768 +trainer/policy/normal/std Min 0.322588 +trainer/policy/normal/log_std Mean 0.994407 +trainer/policy/normal/log_std Std 0.326842 +trainer/policy/normal/log_std Max 1.95695 +trainer/policy/normal/log_std Min -1.13138 +eval/num steps total 973446 +eval/num paths total 976 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0339068 +eval/Actions Std 0.902666 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78439 +time/logging (s) 0.00413544 +time/sampling batch (s) 0.785792 +time/saving (s) 0.00396871 +time/training (s) 7.47743 +time/epoch (s) 11.0557 +time/total (s) 10111 +Epoch -25 +---------------------------------- --------------- +2022-05-10 15:59:34.144405 PDT | [1] Epoch -24 finished +---------------------------------- --------------- +epoch -24 +replay_buffer/size 999033 +trainer/num train calls 977000 +trainer/Policy Loss -20.2938 +trainer/Log Pis Mean 24.0716 +trainer/Log Pis Std 13.2243 +trainer/Log Pis Max 74.8705 +trainer/Log Pis Min -6.00482 +trainer/policy/mean Mean -0.0356374 +trainer/policy/mean Std 0.907738 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76301 +trainer/policy/normal/std Std 0.664636 +trainer/policy/normal/std Max 5.95433 +trainer/policy/normal/std Min 0.287123 +trainer/policy/normal/log_std Mean 0.975865 +trainer/policy/normal/log_std Std 0.317963 +trainer/policy/normal/log_std Max 1.78412 +trainer/policy/normal/log_std Min -1.24784 +eval/num steps total 974446 +eval/num paths total 977 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.314902 +eval/Actions Std 0.847037 +eval/Actions Max 0.999984 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67372 +time/logging (s) 0.00376385 +time/sampling batch (s) 0.281249 +time/saving (s) 0.00357656 +time/training (s) 7.56896 +time/epoch (s) 10.5313 +time/total (s) 10121.6 +Epoch -24 +---------------------------------- --------------- +2022-05-10 15:59:44.526144 PDT | [1] Epoch -23 finished +---------------------------------- --------------- +epoch -23 +replay_buffer/size 999033 +trainer/num train calls 978000 +trainer/Policy Loss -19.9995 +trainer/Log Pis Mean 23.6117 +trainer/Log Pis Std 12.5131 +trainer/Log Pis Max 76.8242 +trainer/Log Pis Min -7.73966 +trainer/policy/mean Mean -0.034376 +trainer/policy/mean Std 0.907972 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.76026 +trainer/policy/normal/std Std 0.681382 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.251354 +trainer/policy/normal/log_std Mean 0.97131 +trainer/policy/normal/log_std Std 0.335539 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.38089 +eval/num steps total 975446 +eval/num paths total 978 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108306 +eval/Actions Std 0.891985 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78869 +time/logging (s) 0.00380297 +time/sampling batch (s) 0.527985 +time/saving (s) 0.00344458 +time/training (s) 7.03228 +time/epoch (s) 10.3562 +time/total (s) 10131.9 +Epoch -23 +---------------------------------- --------------- +2022-05-10 15:59:54.701292 PDT | [1] Epoch -22 finished +---------------------------------- --------------- +epoch -22 +replay_buffer/size 999033 +trainer/num train calls 979000 +trainer/Policy Loss -18.4309 +trainer/Log Pis Mean 23.6342 +trainer/Log Pis Std 13.7963 +trainer/Log Pis Max 70.6133 +trainer/Log Pis Min -8.03447 +trainer/policy/mean Mean -0.0114361 +trainer/policy/mean Std 0.90535 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79357 +trainer/policy/normal/std Std 0.6861 +trainer/policy/normal/std Max 5.7684 +trainer/policy/normal/std Min 0.256108 +trainer/policy/normal/log_std Mean 0.984008 +trainer/policy/normal/log_std Std 0.332094 +trainer/policy/normal/log_std Max 1.7524 +trainer/policy/normal/log_std Min -1.36216 +eval/num steps total 976446 +eval/num paths total 979 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00859233 +eval/Actions Std 0.93137 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76628 +time/logging (s) 0.00381352 +time/sampling batch (s) 0.277025 +time/saving (s) 0.00345444 +time/training (s) 7.09921 +time/epoch (s) 10.1498 +time/total (s) 10142.1 +Epoch -22 +---------------------------------- --------------- +2022-05-10 16:00:05.044918 PDT | [1] Epoch -21 finished +---------------------------------- --------------- +epoch -21 +replay_buffer/size 999033 +trainer/num train calls 980000 +trainer/Policy Loss -20.7844 +trainer/Log Pis Mean 24.1138 +trainer/Log Pis Std 13.1145 +trainer/Log Pis Max 68.8512 +trainer/Log Pis Min -4.27976 +trainer/policy/mean Mean -0.0159188 +trainer/policy/mean Std 0.909858 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78338 +trainer/policy/normal/std Std 0.681424 +trainer/policy/normal/std Max 6.38594 +trainer/policy/normal/std Min 0.181531 +trainer/policy/normal/log_std Mean 0.976976 +trainer/policy/normal/log_std Std 0.353357 +trainer/policy/normal/log_std Max 1.8541 +trainer/policy/normal/log_std Min -1.70633 +eval/num steps total 977446 +eval/num paths total 980 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.15239 +eval/Actions Std 0.807193 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49851 +time/logging (s) 0.00440991 +time/sampling batch (s) 0.283857 +time/saving (s) 0.00442438 +time/training (s) 7.52725 +time/epoch (s) 10.3185 +time/total (s) 10152.4 +Epoch -21 +---------------------------------- --------------- +2022-05-10 16:00:14.806093 PDT | [1] Epoch -20 finished +---------------------------------- --------------- +epoch -20 +replay_buffer/size 999033 +trainer/num train calls 981000 +trainer/Policy Loss -19.8321 +trainer/Log Pis Mean 24.7399 +trainer/Log Pis Std 12.8582 +trainer/Log Pis Max 70.7168 +trainer/Log Pis Min -5.19405 +trainer/policy/mean Mean -0.0332457 +trainer/policy/mean Std 0.906039 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80708 +trainer/policy/normal/std Std 0.690259 +trainer/policy/normal/std Max 6.21108 +trainer/policy/normal/std Min 0.299452 +trainer/policy/normal/log_std Mean 0.989338 +trainer/policy/normal/log_std Std 0.329506 +trainer/policy/normal/log_std Max 1.82633 +trainer/policy/normal/log_std Min -1.2058 +eval/num steps total 978446 +eval/num paths total 981 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.010946 +eval/Actions Std 0.911531 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69723 +time/logging (s) 0.00451725 +time/sampling batch (s) 0.302224 +time/saving (s) 0.00441454 +time/training (s) 6.72524 +time/epoch (s) 9.73362 +time/total (s) 10162.1 +Epoch -20 +---------------------------------- --------------- +2022-05-10 16:00:25.416732 PDT | [1] Epoch -19 finished +---------------------------------- --------------- +epoch -19 +replay_buffer/size 999033 +trainer/num train calls 982000 +trainer/Policy Loss -19.7733 +trainer/Log Pis Mean 25.5566 +trainer/Log Pis Std 14.1164 +trainer/Log Pis Max 64.553 +trainer/Log Pis Min -9.31073 +trainer/policy/mean Mean -0.0302982 +trainer/policy/mean Std 0.909216 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.8428 +trainer/policy/normal/std Std 0.688864 +trainer/policy/normal/std Max 6.02529 +trainer/policy/normal/std Min 0.328542 +trainer/policy/normal/log_std Mean 1.00275 +trainer/policy/normal/log_std Std 0.326333 +trainer/policy/normal/log_std Max 1.79596 +trainer/policy/normal/log_std Min -1.11309 +eval/num steps total 979446 +eval/num paths total 982 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.199526 +eval/Actions Std 0.899973 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77906 +time/logging (s) 0.00452541 +time/sampling batch (s) 0.545596 +time/saving (s) 0.00466115 +time/training (s) 7.249 +time/epoch (s) 10.5828 +time/total (s) 10172.7 +Epoch -19 +---------------------------------- --------------- +2022-05-10 16:00:36.471157 PDT | [1] Epoch -18 finished +---------------------------------- --------------- +epoch -18 +replay_buffer/size 999033 +trainer/num train calls 983000 +trainer/Policy Loss -19.6733 +trainer/Log Pis Mean 24.0005 +trainer/Log Pis Std 13.7216 +trainer/Log Pis Max 68.0992 +trainer/Log Pis Min -8.58148 +trainer/policy/mean Mean -0.0466705 +trainer/policy/mean Std 0.907949 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.74597 +trainer/policy/normal/std Std 0.664539 +trainer/policy/normal/std Max 5.43 +trainer/policy/normal/std Min 0.274261 +trainer/policy/normal/log_std Mean 0.967081 +trainer/policy/normal/log_std Std 0.333707 +trainer/policy/normal/log_std Max 1.69194 +trainer/policy/normal/log_std Min -1.29367 +eval/num steps total 980446 +eval/num paths total 983 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.169355 +eval/Actions Std 0.901079 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73332 +time/logging (s) 0.00438421 +time/sampling batch (s) 0.284208 +time/saving (s) 0.00430104 +time/training (s) 8.00055 +time/epoch (s) 11.0268 +time/total (s) 10183.8 +Epoch -18 +---------------------------------- --------------- +2022-05-10 16:00:46.822368 PDT | [1] Epoch -17 finished +---------------------------------- --------------- +epoch -17 +replay_buffer/size 999033 +trainer/num train calls 984000 +trainer/Policy Loss -19.0142 +trainer/Log Pis Mean 23.7834 +trainer/Log Pis Std 12.3529 +trainer/Log Pis Max 72.4512 +trainer/Log Pis Min -10.5561 +trainer/policy/mean Mean -0.0408347 +trainer/policy/mean Std 0.903978 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.68651 +trainer/policy/normal/std Std 0.654461 +trainer/policy/normal/std Max 5.8208 +trainer/policy/normal/std Min 0.291248 +trainer/policy/normal/log_std Mean 0.944926 +trainer/policy/normal/log_std Std 0.334 +trainer/policy/normal/log_std Max 1.76144 +trainer/policy/normal/log_std Min -1.23358 +eval/num steps total 981446 +eval/num paths total 984 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0447888 +eval/Actions Std 0.896118 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57085 +time/logging (s) 0.0041009 +time/sampling batch (s) 0.542932 +time/saving (s) 0.00415558 +time/training (s) 7.20174 +time/epoch (s) 10.3238 +time/total (s) 10194.1 +Epoch -17 +---------------------------------- --------------- +2022-05-10 16:00:57.299131 PDT | [1] Epoch -16 finished +---------------------------------- --------------- +epoch -16 +replay_buffer/size 999033 +trainer/num train calls 985000 +trainer/Policy Loss -18.9255 +trainer/Log Pis Mean 23.6434 +trainer/Log Pis Std 12.8666 +trainer/Log Pis Max 82.6827 +trainer/Log Pis Min -7.90099 +trainer/policy/mean Mean -0.0172019 +trainer/policy/mean Std 0.908947 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81386 +trainer/policy/normal/std Std 0.670199 +trainer/policy/normal/std Max 6.1523 +trainer/policy/normal/std Min 0.284816 +trainer/policy/normal/log_std Mean 0.99196 +trainer/policy/normal/log_std Std 0.334288 +trainer/policy/normal/log_std Max 1.81683 +trainer/policy/normal/log_std Min -1.25591 +eval/num steps total 982446 +eval/num paths total 985 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113665 +eval/Actions Std 0.882351 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68891 +time/logging (s) 0.00515532 +time/sampling batch (s) 0.291967 +time/saving (s) 0.0058097 +time/training (s) 7.45914 +time/epoch (s) 10.451 +time/total (s) 10204.5 +Epoch -16 +---------------------------------- --------------- +2022-05-10 16:01:07.613301 PDT | [1] Epoch -15 finished +---------------------------------- --------------- +epoch -15 +replay_buffer/size 999033 +trainer/num train calls 986000 +trainer/Policy Loss -20.1589 +trainer/Log Pis Mean 24.1483 +trainer/Log Pis Std 13.5737 +trainer/Log Pis Max 70.6191 +trainer/Log Pis Min -6.75443 +trainer/policy/mean Mean -0.0243759 +trainer/policy/mean Std 0.907317 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76763 +trainer/policy/normal/std Std 0.674026 +trainer/policy/normal/std Max 6.17375 +trainer/policy/normal/std Min 0.182797 +trainer/policy/normal/log_std Mean 0.974774 +trainer/policy/normal/log_std Std 0.334775 +trainer/policy/normal/log_std Max 1.82031 +trainer/policy/normal/log_std Min -1.69938 +eval/num steps total 983446 +eval/num paths total 986 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0141181 +eval/Actions Std 0.902245 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42027 +time/logging (s) 0.00455067 +time/sampling batch (s) 0.291234 +time/saving (s) 0.0045213 +time/training (s) 7.56607 +time/epoch (s) 10.2866 +time/total (s) 10214.8 +Epoch -15 +---------------------------------- --------------- +2022-05-10 16:01:17.593752 PDT | [1] Epoch -14 finished +---------------------------------- --------------- +epoch -14 +replay_buffer/size 999033 +trainer/num train calls 987000 +trainer/Policy Loss -20.9107 +trainer/Log Pis Mean 24.6728 +trainer/Log Pis Std 13.1519 +trainer/Log Pis Max 68.4604 +trainer/Log Pis Min -5.17897 +trainer/policy/mean Mean -0.0438582 +trainer/policy/mean Std 0.908737 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.77946 +trainer/policy/normal/std Std 0.66983 +trainer/policy/normal/std Max 5.92132 +trainer/policy/normal/std Min 0.275123 +trainer/policy/normal/log_std Mean 0.979417 +trainer/policy/normal/log_std Std 0.3326 +trainer/policy/normal/log_std Max 1.77856 +trainer/policy/normal/log_std Min -1.29054 +eval/num steps total 984446 +eval/num paths total 987 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100037 +eval/Actions Std 0.918745 +eval/Actions Max 0.999996 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40379 +time/logging (s) 0.00420294 +time/sampling batch (s) 0.280034 +time/saving (s) 0.00393451 +time/training (s) 7.26187 +time/epoch (s) 9.95383 +time/total (s) 10224.8 +Epoch -14 +---------------------------------- --------------- +2022-05-10 16:01:27.043844 PDT | [1] Epoch -13 finished +---------------------------------- --------------- +epoch -13 +replay_buffer/size 999033 +trainer/num train calls 988000 +trainer/Policy Loss -19.4438 +trainer/Log Pis Mean 24.3205 +trainer/Log Pis Std 13.559 +trainer/Log Pis Max 74.5447 +trainer/Log Pis Min -11.2695 +trainer/policy/mean Mean -0.0306699 +trainer/policy/mean Std 0.902803 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.71394 +trainer/policy/normal/std Std 0.686986 +trainer/policy/normal/std Max 6.9298 +trainer/policy/normal/std Min 0.206951 +trainer/policy/normal/log_std Mean 0.949909 +trainer/policy/normal/log_std Std 0.357161 +trainer/policy/normal/log_std Max 1.93583 +trainer/policy/normal/log_std Min -1.57527 +eval/num steps total 985446 +eval/num paths total 988 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.094398 +eval/Actions Std 0.895869 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66011 +time/logging (s) 0.00387327 +time/sampling batch (s) 0.27891 +time/saving (s) 0.00361934 +time/training (s) 6.47691 +time/epoch (s) 9.42343 +time/total (s) 10234.2 +Epoch -13 +---------------------------------- --------------- +2022-05-10 16:01:36.765777 PDT | [1] Epoch -12 finished +---------------------------------- --------------- +epoch -12 +replay_buffer/size 999033 +trainer/num train calls 989000 +trainer/Policy Loss -20.5724 +trainer/Log Pis Mean 24.6175 +trainer/Log Pis Std 13.3398 +trainer/Log Pis Max 75.3906 +trainer/Log Pis Min -1.97803 +trainer/policy/mean Mean -0.035168 +trainer/policy/mean Std 0.914982 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7961 +trainer/policy/normal/std Std 0.656259 +trainer/policy/normal/std Max 5.37011 +trainer/policy/normal/std Min 0.297218 +trainer/policy/normal/log_std Mean 0.988994 +trainer/policy/normal/log_std Std 0.315869 +trainer/policy/normal/log_std Max 1.68085 +trainer/policy/normal/log_std Min -1.21329 +eval/num steps total 986446 +eval/num paths total 989 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.150246 +eval/Actions Std 0.866215 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53728 +time/logging (s) 0.00453144 +time/sampling batch (s) 0.528265 +time/saving (s) 0.00394611 +time/training (s) 6.62297 +time/epoch (s) 9.69699 +time/total (s) 10243.9 +Epoch -12 +---------------------------------- --------------- +2022-05-10 16:01:47.344642 PDT | [1] Epoch -11 finished +---------------------------------- --------------- +epoch -11 +replay_buffer/size 999033 +trainer/num train calls 990000 +trainer/Policy Loss -19.4763 +trainer/Log Pis Mean 25.579 +trainer/Log Pis Std 13.1369 +trainer/Log Pis Max 66.3938 +trainer/Log Pis Min -8.04126 +trainer/policy/mean Mean -0.0400149 +trainer/policy/mean Std 0.903968 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.79772 +trainer/policy/normal/std Std 0.715704 +trainer/policy/normal/std Max 6.0423 +trainer/policy/normal/std Min 0.226996 +trainer/policy/normal/log_std Mean 0.978712 +trainer/policy/normal/log_std Std 0.36367 +trainer/policy/normal/log_std Max 1.79878 +trainer/policy/normal/log_std Min -1.48282 +eval/num steps total 987446 +eval/num paths total 990 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00182635 +eval/Actions Std 0.908159 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4642 +time/logging (s) 0.00385928 +time/sampling batch (s) 0.2788 +time/saving (s) 0.00352079 +time/training (s) 7.80173 +time/epoch (s) 10.5521 +time/total (s) 10254.5 +Epoch -11 +---------------------------------- --------------- +2022-05-10 16:01:56.739753 PDT | [1] Epoch -10 finished +---------------------------------- --------------- +epoch -10 +replay_buffer/size 999033 +trainer/num train calls 991000 +trainer/Policy Loss -20.3465 +trainer/Log Pis Mean 24.9778 +trainer/Log Pis Std 13.6655 +trainer/Log Pis Max 63.0444 +trainer/Log Pis Min -12.3078 +trainer/policy/mean Mean -0.04666 +trainer/policy/mean Std 0.905524 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76601 +trainer/policy/normal/std Std 0.674657 +trainer/policy/normal/std Max 5.16221 +trainer/policy/normal/std Min 0.256644 +trainer/policy/normal/log_std Mean 0.972278 +trainer/policy/normal/log_std Std 0.344503 +trainer/policy/normal/log_std Max 1.64136 +trainer/policy/normal/log_std Min -1.36007 +eval/num steps total 988446 +eval/num paths total 991 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.181733 +eval/Actions Std 0.892028 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45947 +time/logging (s) 0.00423624 +time/sampling batch (s) 0.279692 +time/saving (s) 0.00381964 +time/training (s) 6.62247 +time/epoch (s) 9.36968 +time/total (s) 10263.8 +Epoch -10 +---------------------------------- --------------- +2022-05-10 16:02:06.583963 PDT | [1] Epoch -9 finished +---------------------------------- --------------- +epoch -9 +replay_buffer/size 999033 +trainer/num train calls 992000 +trainer/Policy Loss -20.1042 +trainer/Log Pis Mean 23.9572 +trainer/Log Pis Std 12.5055 +trainer/Log Pis Max 74.4304 +trainer/Log Pis Min -10.447 +trainer/policy/mean Mean -0.0616889 +trainer/policy/mean Std 0.902862 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79602 +trainer/policy/normal/std Std 0.704253 +trainer/policy/normal/std Max 7.32627 +trainer/policy/normal/std Min 0.257121 +trainer/policy/normal/log_std Mean 0.982017 +trainer/policy/normal/log_std Std 0.345728 +trainer/policy/normal/log_std Max 1.99147 +trainer/policy/normal/log_std Min -1.35821 +eval/num steps total 989446 +eval/num paths total 992 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.054607 +eval/Actions Std 0.902829 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.656 +time/logging (s) 0.00410897 +time/sampling batch (s) 0.284869 +time/saving (s) 0.00379511 +time/training (s) 6.86873 +time/epoch (s) 9.81751 +time/total (s) 10273.7 +Epoch -9 +---------------------------------- --------------- +2022-05-10 16:02:16.078863 PDT | [1] Epoch -8 finished +---------------------------------- --------------- +epoch -8 +replay_buffer/size 999033 +trainer/num train calls 993000 +trainer/Policy Loss -19.1474 +trainer/Log Pis Mean 25.2795 +trainer/Log Pis Std 13.099 +trainer/Log Pis Max 71.3303 +trainer/Log Pis Min -10.8088 +trainer/policy/mean Mean -0.0506014 +trainer/policy/mean Std 0.903709 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81566 +trainer/policy/normal/std Std 0.693322 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.234868 +trainer/policy/normal/log_std Mean 0.991644 +trainer/policy/normal/log_std Std 0.334708 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.44873 +eval/num steps total 990446 +eval/num paths total 993 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.289214 +eval/Actions Std 0.884777 +eval/Actions Max 0.999974 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55001 +time/logging (s) 0.00383798 +time/sampling batch (s) 0.28192 +time/saving (s) 0.00349243 +time/training (s) 6.62922 +time/epoch (s) 9.46847 +time/total (s) 10283.1 +Epoch -8 +---------------------------------- --------------- +2022-05-10 16:02:26.061680 PDT | [1] Epoch -7 finished +---------------------------------- --------------- +epoch -7 +replay_buffer/size 999033 +trainer/num train calls 994000 +trainer/Policy Loss -18.3901 +trainer/Log Pis Mean 24.8461 +trainer/Log Pis Std 13.0566 +trainer/Log Pis Max 72.0953 +trainer/Log Pis Min -7.32508 +trainer/policy/mean Mean -0.0417423 +trainer/policy/mean Std 0.903363 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.75806 +trainer/policy/normal/std Std 0.714116 +trainer/policy/normal/std Max 6.29061 +trainer/policy/normal/std Min 0.240557 +trainer/policy/normal/log_std Mean 0.964014 +trainer/policy/normal/log_std Std 0.363772 +trainer/policy/normal/log_std Max 1.83906 +trainer/policy/normal/log_std Min -1.4248 +eval/num steps total 991446 +eval/num paths total 994 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.175433 +eval/Actions Std 0.932059 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7634 +time/logging (s) 0.00426444 +time/sampling batch (s) 0.529632 +time/saving (s) 0.00411386 +time/training (s) 6.6561 +time/epoch (s) 9.95751 +time/total (s) 10293.1 +Epoch -7 +---------------------------------- --------------- +2022-05-10 16:02:36.333522 PDT | [1] Epoch -6 finished +---------------------------------- --------------- +epoch -6 +replay_buffer/size 999033 +trainer/num train calls 995000 +trainer/Policy Loss -19.1885 +trainer/Log Pis Mean 24.5743 +trainer/Log Pis Std 13.7932 +trainer/Log Pis Max 72.0392 +trainer/Log Pis Min -10.8971 +trainer/policy/mean Mean -0.0308499 +trainer/policy/mean Std 0.905801 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81838 +trainer/policy/normal/std Std 0.661782 +trainer/policy/normal/std Max 5.58155 +trainer/policy/normal/std Min 0.296963 +trainer/policy/normal/log_std Mean 0.996558 +trainer/policy/normal/log_std Std 0.317887 +trainer/policy/normal/log_std Max 1.71947 +trainer/policy/normal/log_std Min -1.21415 +eval/num steps total 992446 +eval/num paths total 995 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.312575 +eval/Actions Std 0.863344 +eval/Actions Max 0.999988 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68253 +time/logging (s) 0.00379267 +time/sampling batch (s) 0.527083 +time/saving (s) 0.00362089 +time/training (s) 7.02814 +time/epoch (s) 10.2452 +time/total (s) 10303.3 +Epoch -6 +---------------------------------- --------------- +2022-05-10 16:02:45.846018 PDT | [1] Epoch -5 finished +---------------------------------- --------------- +epoch -5 +replay_buffer/size 999033 +trainer/num train calls 996000 +trainer/Policy Loss -20.152 +trainer/Log Pis Mean 25.491 +trainer/Log Pis Std 13.6511 +trainer/Log Pis Max 84.9199 +trainer/Log Pis Min -9.32311 +trainer/policy/mean Mean -0.0259265 +trainer/policy/mean Std 0.9089 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78851 +trainer/policy/normal/std Std 0.689809 +trainer/policy/normal/std Max 5.83956 +trainer/policy/normal/std Min 0.215233 +trainer/policy/normal/log_std Mean 0.981017 +trainer/policy/normal/log_std Std 0.339122 +trainer/policy/normal/log_std Max 1.76466 +trainer/policy/normal/log_std Min -1.53604 +eval/num steps total 993446 +eval/num paths total 996 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.299964 +eval/Actions Std 0.859141 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63266 +time/logging (s) 0.00380109 +time/sampling batch (s) 0.524968 +time/saving (s) 0.00340107 +time/training (s) 6.3223 +time/epoch (s) 9.48713 +time/total (s) 10312.8 +Epoch -5 +---------------------------------- --------------- +2022-05-10 16:02:56.221049 PDT | [1] Epoch -4 finished +---------------------------------- --------------- +epoch -4 +replay_buffer/size 999033 +trainer/num train calls 997000 +trainer/Policy Loss -18.4517 +trainer/Log Pis Mean 24.671 +trainer/Log Pis Std 13.852 +trainer/Log Pis Max 67.7829 +trainer/Log Pis Min -7.89741 +trainer/policy/mean Mean -0.0550636 +trainer/policy/mean Std 0.902154 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.86431 +trainer/policy/normal/std Std 0.706817 +trainer/policy/normal/std Max 5.71841 +trainer/policy/normal/std Min 0.304019 +trainer/policy/normal/log_std Mean 1.00756 +trainer/policy/normal/log_std Std 0.339635 +trainer/policy/normal/log_std Max 1.74369 +trainer/policy/normal/log_std Min -1.19067 +eval/num steps total 994446 +eval/num paths total 997 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.192714 +eval/Actions Std 0.746591 +eval/Actions Max 0.999957 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6718 +time/logging (s) 0.00485163 +time/sampling batch (s) 0.527053 +time/saving (s) 0.00403879 +time/training (s) 7.1429 +time/epoch (s) 10.3506 +time/total (s) 10323.2 +Epoch -4 +---------------------------------- --------------- +2022-05-10 16:03:05.912318 PDT | [1] Epoch -3 finished +---------------------------------- --------------- +epoch -3 +replay_buffer/size 999033 +trainer/num train calls 998000 +trainer/Policy Loss -19.9092 +trainer/Log Pis Mean 24.7259 +trainer/Log Pis Std 13.0027 +trainer/Log Pis Max 67.4058 +trainer/Log Pis Min -14.509 +trainer/policy/mean Mean -0.0428854 +trainer/policy/mean Std 0.910499 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.75343 +trainer/policy/normal/std Std 0.64873 +trainer/policy/normal/std Max 5.1461 +trainer/policy/normal/std Min 0.283549 +trainer/policy/normal/log_std Mean 0.972465 +trainer/policy/normal/log_std Std 0.321592 +trainer/policy/normal/log_std Max 1.63824 +trainer/policy/normal/log_std Min -1.26037 +eval/num steps total 995446 +eval/num paths total 998 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.096696 +eval/Actions Std 0.921709 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74352 +time/logging (s) 0.00409472 +time/sampling batch (s) 0.53595 +time/saving (s) 0.00376078 +time/training (s) 6.37589 +time/epoch (s) 9.66322 +time/total (s) 10332.9 +Epoch -3 +---------------------------------- --------------- +2022-05-10 16:03:13.992880 PDT | [1] Epoch -2 finished +---------------------------------- --------------- +epoch -2 +replay_buffer/size 999033 +trainer/num train calls 999000 +trainer/Policy Loss -19.1104 +trainer/Log Pis Mean 24.2706 +trainer/Log Pis Std 13.1849 +trainer/Log Pis Max 65.5975 +trainer/Log Pis Min -13.0876 +trainer/policy/mean Mean -0.025871 +trainer/policy/mean Std 0.904351 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.76184 +trainer/policy/normal/std Std 0.675454 +trainer/policy/normal/std Max 5.82507 +trainer/policy/normal/std Min 0.287782 +trainer/policy/normal/log_std Mean 0.971268 +trainer/policy/normal/log_std Std 0.339788 +trainer/policy/normal/log_std Max 1.76217 +trainer/policy/normal/log_std Min -1.24555 +eval/num steps total 996446 +eval/num paths total 999 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123443 +eval/Actions Std 0.895815 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 1.9732 +time/logging (s) 0.0040371 +time/sampling batch (s) 0.530343 +time/saving (s) 0.00392802 +time/training (s) 5.54309 +time/epoch (s) 8.0546 +time/total (s) 10340.9 +Epoch -2 +---------------------------------- --------------- +2022-05-10 16:03:20.948625 PDT | [1] Epoch -1 finished +---------------------------------- --------------- +epoch -1 +replay_buffer/size 999033 +trainer/num train calls 1e+06 +trainer/Policy Loss -20.8051 +trainer/Log Pis Mean 24.3261 +trainer/Log Pis Std 13.7959 +trainer/Log Pis Max 72.7107 +trainer/Log Pis Min -8.77733 +trainer/policy/mean Mean -0.041018 +trainer/policy/mean Std 0.908495 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.75361 +trainer/policy/normal/std Std 0.671789 +trainer/policy/normal/std Max 6.01655 +trainer/policy/normal/std Min 0.289692 +trainer/policy/normal/log_std Mean 0.969625 +trainer/policy/normal/log_std Std 0.333882 +trainer/policy/normal/log_std Max 1.79451 +trainer/policy/normal/log_std Min -1.23894 +eval/num steps total 997446 +eval/num paths total 1000 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.43026 +eval/Actions Std 0.693621 +eval/Actions Max 0.999995 +eval/Actions Min -0.999959 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 1.82664 +time/logging (s) 0.0040019 +time/sampling batch (s) 0.277167 +time/saving (s) 0.00683284 +time/training (s) 4.81525 +time/epoch (s) 6.9299 +time/total (s) 10347.8 +Epoch -1 +---------------------------------- ---------------