diff --git "a/sg/antmaze-medium-diverse-v0/0/debug.log" "b/sg/antmaze-medium-diverse-v0/0/debug.log" new file mode 100644--- /dev/null +++ "b/sg/antmaze-medium-diverse-v0/0/debug.log" @@ -0,0 +1,52000 @@ +2022-05-10 13:11:18.239757 PDT | [0] Epoch -1000 finished +---------------------------------- ---------------- +epoch -1000 +replay_buffer/size 999033 +trainer/num train calls 1000 +trainer/Policy Loss 39.1983 +trainer/Log Pis Mean -5.34762 +trainer/Log Pis Std 0.628231 +trainer/Log Pis Max -3.55754 +trainer/Log Pis Min -7.0817 +trainer/policy/mean Mean 4.56278e-05 +trainer/policy/mean Std 0.000163986 +trainer/policy/mean Max 0.000636091 +trainer/policy/mean Min -0.000624096 +trainer/policy/normal/std Mean 0.999732 +trainer/policy/normal/std Std 0.000610854 +trainer/policy/normal/std Max 1.00095 +trainer/policy/normal/std Min 0.998414 +trainer/policy/normal/log_std Mean -0.000268683 +trainer/policy/normal/log_std Std 0.000611058 +trainer/policy/normal/log_std Max 0.000949766 +trainer/policy/normal/log_std Min -0.00158764 +eval/num steps total 1000 +eval/num paths total 1 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 1.79059e-05 +eval/Actions Std 8.27542e-05 +eval/Actions Max 0.000358391 +eval/Actions Min -0.000323405 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 4.57378 +time/logging (s) 0.00436546 +time/sampling batch (s) 0.28522 +time/saving (s) 0.00458879 +time/training (s) 6.0399 +time/epoch (s) 10.9078 +time/total (s) 44.816 +Epoch -1000 +---------------------------------- ---------------- +2022-05-10 13:11:28.288166 PDT | [0] Epoch -999 finished +---------------------------------- --------------- +epoch -999 +replay_buffer/size 999033 +trainer/num train calls 2000 +trainer/Policy Loss -18.0775 +trainer/Log Pis Mean 21.2846 +trainer/Log Pis Std 13.1061 +trainer/Log Pis Max 65.7455 +trainer/Log Pis Min -11.9654 +trainer/policy/mean Mean -0.0534806 +trainer/policy/mean Std 0.731282 +trainer/policy/mean Max 0.998808 +trainer/policy/mean Min -0.998915 +trainer/policy/normal/std Mean 3.92232 +trainer/policy/normal/std Std 0.607491 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 2.19549 +trainer/policy/normal/log_std Mean 1.35481 +trainer/policy/normal/log_std Std 0.15438 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min 0.786407 +eval/num steps total 2000 +eval/num paths total 2 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0559851 +eval/Actions Std 0.70347 +eval/Actions Max 0.982046 +eval/Actions Min -0.985315 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73167 +time/logging (s) 0.00473255 +time/sampling batch (s) 0.279189 +time/saving (s) 0.00483205 +time/training (s) 7.00917 +time/epoch (s) 10.0296 +time/total (s) 54.8493 +Epoch -999 +---------------------------------- --------------- +2022-05-10 13:11:37.796220 PDT | [0] Epoch -998 finished +---------------------------------- --------------- +epoch -998 +replay_buffer/size 999033 +trainer/num train calls 3000 +trainer/Policy Loss -17.3079 +trainer/Log Pis Mean 21.6775 +trainer/Log Pis Std 12.87 +trainer/Log Pis Max 78.127 +trainer/Log Pis Min -5.84775 +trainer/policy/mean Mean -0.062642 +trainer/policy/mean Std 0.834923 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999967 +trainer/policy/normal/std Mean 3.48965 +trainer/policy/normal/std Std 0.480764 +trainer/policy/normal/std Max 5.18909 +trainer/policy/normal/std Min 1.75664 +trainer/policy/normal/log_std Mean 1.23988 +trainer/policy/normal/log_std Std 0.14304 +trainer/policy/normal/log_std Max 1.64656 +trainer/policy/normal/log_std Min 0.563403 +eval/num steps total 3000 +eval/num paths total 3 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0607293 +eval/Actions Std 0.675168 +eval/Actions Max 0.998547 +eval/Actions Min -0.984274 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62037 +time/logging (s) 0.00410002 +time/sampling batch (s) 0.532802 +time/saving (s) 0.00424397 +time/training (s) 6.32655 +time/epoch (s) 9.48807 +time/total (s) 64.3411 +Epoch -998 +---------------------------------- --------------- +2022-05-10 13:11:48.405889 PDT | [0] Epoch -997 finished +---------------------------------- --------------- +epoch -997 +replay_buffer/size 999033 +trainer/num train calls 4000 +trainer/Policy Loss -17.4478 +trainer/Log Pis Mean 22.9257 +trainer/Log Pis Std 12.6806 +trainer/Log Pis Max 67.6542 +trainer/Log Pis Min -3.49838 +trainer/policy/mean Mean -0.0647393 +trainer/policy/mean Std 0.852645 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 3.32607 +trainer/policy/normal/std Std 0.428262 +trainer/policy/normal/std Max 5.02023 +trainer/policy/normal/std Min 1.59782 +trainer/policy/normal/log_std Mean 1.19307 +trainer/policy/normal/log_std Std 0.134505 +trainer/policy/normal/log_std Max 1.61348 +trainer/policy/normal/log_std Min 0.468639 +eval/num steps total 4000 +eval/num paths total 4 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00441327 +eval/Actions Std 0.790933 +eval/Actions Max 0.999861 +eval/Actions Min -0.999938 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79814 +time/logging (s) 0.00378746 +time/sampling batch (s) 0.286242 +time/saving (s) 0.00363749 +time/training (s) 7.49842 +time/epoch (s) 10.5902 +time/total (s) 74.9347 +Epoch -997 +---------------------------------- --------------- +2022-05-10 13:11:59.922664 PDT | [0] Epoch -996 finished +---------------------------------- --------------- +epoch -996 +replay_buffer/size 999033 +trainer/num train calls 5000 +trainer/Policy Loss -17.4005 +trainer/Log Pis Mean 23.5806 +trainer/Log Pis Std 12.6158 +trainer/Log Pis Max 74.2838 +trainer/Log Pis Min -5.90942 +trainer/policy/mean Mean -0.0508461 +trainer/policy/mean Std 0.868638 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.20556 +trainer/policy/normal/std Std 0.449807 +trainer/policy/normal/std Max 5.34089 +trainer/policy/normal/std Min 1.42882 +trainer/policy/normal/log_std Mean 1.15447 +trainer/policy/normal/log_std Std 0.147571 +trainer/policy/normal/log_std Max 1.67539 +trainer/policy/normal/log_std Min 0.356852 +eval/num steps total 5000 +eval/num paths total 5 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.281674 +eval/Actions Std 0.851804 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56461 +time/logging (s) 0.00406958 +time/sampling batch (s) 0.286236 +time/saving (s) 0.0041653 +time/training (s) 8.63932 +time/epoch (s) 11.4984 +time/total (s) 86.436 +Epoch -996 +---------------------------------- --------------- +2022-05-10 13:12:11.297127 PDT | [0] Epoch -995 finished +---------------------------------- --------------- +epoch -995 +replay_buffer/size 999033 +trainer/num train calls 6000 +trainer/Policy Loss -19.0485 +trainer/Log Pis Mean 23.1984 +trainer/Log Pis Std 13.1019 +trainer/Log Pis Max 64.4764 +trainer/Log Pis Min -8.44971 +trainer/policy/mean Mean -0.0330821 +trainer/policy/mean Std 0.879392 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.17695 +trainer/policy/normal/std Std 0.409103 +trainer/policy/normal/std Max 5.43729 +trainer/policy/normal/std Min 1.33026 +trainer/policy/normal/log_std Mean 1.1469 +trainer/policy/normal/log_std Std 0.138445 +trainer/policy/normal/log_std Max 1.69328 +trainer/policy/normal/log_std Min 0.285374 +eval/num steps total 6000 +eval/num paths total 6 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.240779 +eval/Actions Std 0.849647 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79243 +time/logging (s) 0.00414192 +time/sampling batch (s) 0.285754 +time/saving (s) 0.0043178 +time/training (s) 8.26892 +time/epoch (s) 11.3556 +time/total (s) 97.7948 +Epoch -995 +---------------------------------- --------------- +2022-05-10 13:12:22.154590 PDT | [0] Epoch -994 finished +---------------------------------- ---------------- +epoch -994 +replay_buffer/size 999033 +trainer/num train calls 7000 +trainer/Policy Loss -17.5005 +trainer/Log Pis Mean 24.5704 +trainer/Log Pis Std 13.2173 +trainer/Log Pis Max 73.5443 +trainer/Log Pis Min -7.3254 +trainer/policy/mean Mean 0.000292083 +trainer/policy/mean Std 0.885989 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.13968 +trainer/policy/normal/std Std 0.401967 +trainer/policy/normal/std Max 5.07576 +trainer/policy/normal/std Min 1.17523 +trainer/policy/normal/log_std Mean 1.13512 +trainer/policy/normal/log_std Std 0.138461 +trainer/policy/normal/log_std Max 1.62448 +trainer/policy/normal/log_std Min 0.161462 +eval/num steps total 7000 +eval/num paths total 7 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0171951 +eval/Actions Std 0.860895 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59823 +time/logging (s) 0.00417366 +time/sampling batch (s) 1.03802 +time/saving (s) 0.0043448 +time/training (s) 7.19334 +time/epoch (s) 10.8381 +time/total (s) 108.636 +Epoch -994 +---------------------------------- ---------------- +2022-05-10 13:12:33.121698 PDT | [0] Epoch -993 finished +---------------------------------- --------------- +epoch -993 +replay_buffer/size 999033 +trainer/num train calls 8000 +trainer/Policy Loss -19.2405 +trainer/Log Pis Mean 23.6208 +trainer/Log Pis Std 12.645 +trainer/Log Pis Max 61.3762 +trainer/Log Pis Min -11.8309 +trainer/policy/mean Mean -0.0271021 +trainer/policy/mean Std 0.89064 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 3.07041 +trainer/policy/normal/std Std 0.43886 +trainer/policy/normal/std Max 5.25414 +trainer/policy/normal/std Min 1.15332 +trainer/policy/normal/log_std Mean 1.11033 +trainer/policy/normal/log_std Std 0.157673 +trainer/policy/normal/log_std Max 1.65902 +trainer/policy/normal/log_std Min 0.142645 +eval/num steps total 8000 +eval/num paths total 8 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0410582 +eval/Actions Std 0.880241 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5276 +time/logging (s) 0.00374276 +time/sampling batch (s) 0.534554 +time/saving (s) 0.00357359 +time/training (s) 7.87814 +time/epoch (s) 10.9476 +time/total (s) 119.587 +Epoch -993 +---------------------------------- --------------- +2022-05-10 13:12:45.230135 PDT | [0] Epoch -992 finished +---------------------------------- --------------- +epoch -992 +replay_buffer/size 999033 +trainer/num train calls 9000 +trainer/Policy Loss -17.2068 +trainer/Log Pis Mean 22.9004 +trainer/Log Pis Std 12.8133 +trainer/Log Pis Max 68.9865 +trainer/Log Pis Min -6.78638 +trainer/policy/mean Mean -0.0455542 +trainer/policy/mean Std 0.888436 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 3.07532 +trainer/policy/normal/std Std 0.419349 +trainer/policy/normal/std Max 4.59009 +trainer/policy/normal/std Min 1.07362 +trainer/policy/normal/log_std Mean 1.11261 +trainer/policy/normal/log_std Std 0.153977 +trainer/policy/normal/log_std Max 1.5239 +trainer/policy/normal/log_std Min 0.071036 +eval/num steps total 9000 +eval/num paths total 9 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109488 +eval/Actions Std 0.880475 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58397 +time/logging (s) 0.0040135 +time/sampling batch (s) 0.534355 +time/saving (s) 0.00410596 +time/training (s) 8.96382 +time/epoch (s) 12.0903 +time/total (s) 131.681 +Epoch -992 +---------------------------------- --------------- +2022-05-10 13:12:55.790694 PDT | [0] Epoch -991 finished +---------------------------------- --------------- +epoch -991 +replay_buffer/size 999033 +trainer/num train calls 10000 +trainer/Policy Loss -19.4214 +trainer/Log Pis Mean 24.3306 +trainer/Log Pis Std 12.8987 +trainer/Log Pis Max 77.9897 +trainer/Log Pis Min -10.5793 +trainer/policy/mean Mean -0.0513071 +trainer/policy/mean Std 0.888829 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 3.08365 +trainer/policy/normal/std Std 0.429094 +trainer/policy/normal/std Max 4.7845 +trainer/policy/normal/std Min 1.03168 +trainer/policy/normal/log_std Mean 1.11474 +trainer/policy/normal/log_std Std 0.158783 +trainer/policy/normal/log_std Max 1.56538 +trainer/policy/normal/log_std Min 0.0311839 +eval/num steps total 10000 +eval/num paths total 10 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.306836 +eval/Actions Std 0.819771 +eval/Actions Max 0.999986 +eval/Actions Min -0.999912 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56315 +time/logging (s) 0.00415608 +time/sampling batch (s) 0.533473 +time/saving (s) 0.00417361 +time/training (s) 7.43694 +time/epoch (s) 10.5419 +time/total (s) 142.226 +Epoch -991 +---------------------------------- --------------- +2022-05-10 13:13:05.740425 PDT | [0] Epoch -990 finished +---------------------------------- --------------- +epoch -990 +replay_buffer/size 999033 +trainer/num train calls 11000 +trainer/Policy Loss -18.4663 +trainer/Log Pis Mean 22.792 +trainer/Log Pis Std 12.0428 +trainer/Log Pis Max 70.0962 +trainer/Log Pis Min -6.61091 +trainer/policy/mean Mean -0.0180358 +trainer/policy/mean Std 0.890227 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 3.08436 +trainer/policy/normal/std Std 0.438512 +trainer/policy/normal/std Max 5.27862 +trainer/policy/normal/std Min 0.991958 +trainer/policy/normal/log_std Mean 1.11475 +trainer/policy/normal/log_std Std 0.159304 +trainer/policy/normal/log_std Max 1.66367 +trainer/policy/normal/log_std Min -0.00807479 +eval/num steps total 11000 +eval/num paths total 11 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107721 +eval/Actions Std 0.873787 +eval/Actions Max 0.99998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5812 +time/logging (s) 0.0038617 +time/sampling batch (s) 0.280288 +time/saving (s) 0.00363731 +time/training (s) 7.06148 +time/epoch (s) 9.93046 +time/total (s) 152.16 +Epoch -990 +---------------------------------- --------------- +2022-05-10 13:13:16.014682 PDT | [0] Epoch -989 finished +---------------------------------- --------------- +epoch -989 +replay_buffer/size 999033 +trainer/num train calls 12000 +trainer/Policy Loss -18.8044 +trainer/Log Pis Mean 23.7012 +trainer/Log Pis Std 13.1894 +trainer/Log Pis Max 73.3975 +trainer/Log Pis Min -5.24693 +trainer/policy/mean Mean -0.0278906 +trainer/policy/mean Std 0.890919 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.06803 +trainer/policy/normal/std Std 0.432118 +trainer/policy/normal/std Max 4.76221 +trainer/policy/normal/std Min 0.995705 +trainer/policy/normal/log_std Mean 1.10923 +trainer/policy/normal/log_std Std 0.16245 +trainer/policy/normal/log_std Max 1.56071 +trainer/policy/normal/log_std Min -0.00430406 +eval/num steps total 12000 +eval/num paths total 12 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00501772 +eval/Actions Std 0.924722 +eval/Actions Max 0.999994 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58343 +time/logging (s) 0.00371093 +time/sampling batch (s) 0.530285 +time/saving (s) 0.00355751 +time/training (s) 7.13478 +time/epoch (s) 10.2558 +time/total (s) 162.418 +Epoch -989 +---------------------------------- --------------- +2022-05-10 13:13:26.171694 PDT | [0] Epoch -988 finished +---------------------------------- --------------- +epoch -988 +replay_buffer/size 999033 +trainer/num train calls 13000 +trainer/Policy Loss -19.2151 +trainer/Log Pis Mean 23.7632 +trainer/Log Pis Std 12.6851 +trainer/Log Pis Max 74.8025 +trainer/Log Pis Min -7.64682 +trainer/policy/mean Mean -0.0393202 +trainer/policy/mean Std 0.893801 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.99002 +trainer/policy/normal/std Std 0.439601 +trainer/policy/normal/std Max 5.19904 +trainer/policy/normal/std Min 0.918162 +trainer/policy/normal/log_std Mean 1.08222 +trainer/policy/normal/log_std Std 0.17179 +trainer/policy/normal/log_std Max 1.64847 +trainer/policy/normal/log_std Min -0.0853812 +eval/num steps total 13000 +eval/num paths total 13 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118019 +eval/Actions Std 0.885746 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62293 +time/logging (s) 0.00425366 +time/sampling batch (s) 0.530506 +time/saving (s) 0.0040948 +time/training (s) 6.9774 +time/epoch (s) 10.1392 +time/total (s) 172.56 +Epoch -988 +---------------------------------- --------------- +2022-05-10 13:13:37.524854 PDT | [0] Epoch -987 finished +---------------------------------- --------------- +epoch -987 +replay_buffer/size 999033 +trainer/num train calls 14000 +trainer/Policy Loss -17.9284 +trainer/Log Pis Mean 24.523 +trainer/Log Pis Std 13.1752 +trainer/Log Pis Max 59.7401 +trainer/Log Pis Min -7.07867 +trainer/policy/mean Mean -0.0443041 +trainer/policy/mean Std 0.895346 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 3.02574 +trainer/policy/normal/std Std 0.448291 +trainer/policy/normal/std Max 5.24538 +trainer/policy/normal/std Min 0.863173 +trainer/policy/normal/log_std Mean 1.09394 +trainer/policy/normal/log_std Std 0.17276 +trainer/policy/normal/log_std Max 1.65735 +trainer/policy/normal/log_std Min -0.14714 +eval/num steps total 14000 +eval/num paths total 14 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117137 +eval/Actions Std 0.876417 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81025 +time/logging (s) 0.00401758 +time/sampling batch (s) 0.284829 +time/saving (s) 0.00364105 +time/training (s) 8.23145 +time/epoch (s) 11.3342 +time/total (s) 183.898 +Epoch -987 +---------------------------------- --------------- +2022-05-10 13:13:48.119307 PDT | [0] Epoch -986 finished +---------------------------------- --------------- +epoch -986 +replay_buffer/size 999033 +trainer/num train calls 15000 +trainer/Policy Loss -19.3749 +trainer/Log Pis Mean 24.5276 +trainer/Log Pis Std 12.3454 +trainer/Log Pis Max 64.0052 +trainer/Log Pis Min -6.09945 +trainer/policy/mean Mean -0.048289 +trainer/policy/mean Std 0.896345 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.96807 +trainer/policy/normal/std Std 0.44893 +trainer/policy/normal/std Max 5.32032 +trainer/policy/normal/std Min 0.859673 +trainer/policy/normal/log_std Mean 1.0742 +trainer/policy/normal/log_std Std 0.175881 +trainer/policy/normal/log_std Max 1.67153 +trainer/policy/normal/log_std Min -0.151203 +eval/num steps total 15000 +eval/num paths total 15 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.01639 +eval/Actions Std 0.863262 +eval/Actions Max 0.999998 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5168 +time/logging (s) 0.00374126 +time/sampling batch (s) 0.27927 +time/saving (s) 0.00349002 +time/training (s) 7.7723 +time/epoch (s) 10.5756 +time/total (s) 194.476 +Epoch -986 +---------------------------------- --------------- +2022-05-10 13:13:59.073082 PDT | [0] Epoch -985 finished +---------------------------------- --------------- +epoch -985 +replay_buffer/size 999033 +trainer/num train calls 16000 +trainer/Policy Loss -19.0231 +trainer/Log Pis Mean 23.3431 +trainer/Log Pis Std 12.2387 +trainer/Log Pis Max 62.3495 +trainer/Log Pis Min -6.93784 +trainer/policy/mean Mean -0.0398874 +trainer/policy/mean Std 0.893742 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.99433 +trainer/policy/normal/std Std 0.457065 +trainer/policy/normal/std Max 5.31152 +trainer/policy/normal/std Min 0.94342 +trainer/policy/normal/log_std Mean 1.08292 +trainer/policy/normal/log_std Std 0.175662 +trainer/policy/normal/log_std Max 1.66988 +trainer/policy/normal/log_std Min -0.0582436 +eval/num steps total 16000 +eval/num paths total 16 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.42536 +eval/Actions Std 0.881444 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58396 +time/logging (s) 0.00419445 +time/sampling batch (s) 0.282582 +time/saving (s) 0.00408483 +time/training (s) 8.06071 +time/epoch (s) 10.9355 +time/total (s) 205.415 +Epoch -985 +---------------------------------- --------------- +2022-05-10 13:14:10.744910 PDT | [0] Epoch -984 finished +---------------------------------- --------------- +epoch -984 +replay_buffer/size 999033 +trainer/num train calls 17000 +trainer/Policy Loss -19.845 +trainer/Log Pis Mean 24.3185 +trainer/Log Pis Std 12.5825 +trainer/Log Pis Max 70.608 +trainer/Log Pis Min -7.08322 +trainer/policy/mean Mean -0.0553788 +trainer/policy/mean Std 0.897198 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.98714 +trainer/policy/normal/std Std 0.429103 +trainer/policy/normal/std Max 4.75471 +trainer/policy/normal/std Min 0.884062 +trainer/policy/normal/log_std Mean 1.08172 +trainer/policy/normal/log_std Std 0.169074 +trainer/policy/normal/log_std Max 1.55914 +trainer/policy/normal/log_std Min -0.123228 +eval/num steps total 17000 +eval/num paths total 17 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.12559 +eval/Actions Std 0.892977 +eval/Actions Max 0.999992 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65235 +time/logging (s) 0.00374059 +time/sampling batch (s) 0.535728 +time/saving (s) 0.00366505 +time/training (s) 8.45694 +time/epoch (s) 11.6524 +time/total (s) 217.071 +Epoch -984 +---------------------------------- --------------- +2022-05-10 13:14:20.063028 PDT | [0] Epoch -983 finished +---------------------------------- --------------- +epoch -983 +replay_buffer/size 999033 +trainer/num train calls 18000 +trainer/Policy Loss -18.9852 +trainer/Log Pis Mean 23.8134 +trainer/Log Pis Std 13.3716 +trainer/Log Pis Max 67.926 +trainer/Log Pis Min -8.40725 +trainer/policy/mean Mean -0.0368374 +trainer/policy/mean Std 0.896538 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.96294 +trainer/policy/normal/std Std 0.472507 +trainer/policy/normal/std Max 5.47726 +trainer/policy/normal/std Min 0.843847 +trainer/policy/normal/log_std Mean 1.07092 +trainer/policy/normal/log_std Std 0.185127 +trainer/policy/normal/log_std Max 1.7006 +trainer/policy/normal/log_std Min -0.169784 +eval/num steps total 18000 +eval/num paths total 18 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0255129 +eval/Actions Std 0.903422 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63608 +time/logging (s) 0.00365618 +time/sampling batch (s) 0.283623 +time/saving (s) 0.00350378 +time/training (s) 6.37243 +time/epoch (s) 9.2993 +time/total (s) 226.373 +Epoch -983 +---------------------------------- --------------- +2022-05-10 13:14:29.998286 PDT | [0] Epoch -982 finished +---------------------------------- --------------- +epoch -982 +replay_buffer/size 999033 +trainer/num train calls 19000 +trainer/Policy Loss -17.9594 +trainer/Log Pis Mean 24.0172 +trainer/Log Pis Std 12.2369 +trainer/Log Pis Max 58.9114 +trainer/Log Pis Min -6.34454 +trainer/policy/mean Mean -0.0360333 +trainer/policy/mean Std 0.891492 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 3.01072 +trainer/policy/normal/std Std 0.477552 +trainer/policy/normal/std Max 5.35421 +trainer/policy/normal/std Min 0.97823 +trainer/policy/normal/log_std Mean 1.08693 +trainer/policy/normal/log_std Std 0.185482 +trainer/policy/normal/log_std Max 1.67788 +trainer/policy/normal/log_std Min -0.02201 +eval/num steps total 19000 +eval/num paths total 19 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0353396 +eval/Actions Std 0.881274 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75398 +time/logging (s) 0.00402175 +time/sampling batch (s) 0.288119 +time/saving (s) 0.00414785 +time/training (s) 6.86636 +time/epoch (s) 9.91663 +time/total (s) 236.292 +Epoch -982 +---------------------------------- --------------- +2022-05-10 13:14:39.699931 PDT | [0] Epoch -981 finished +---------------------------------- --------------- +epoch -981 +replay_buffer/size 999033 +trainer/num train calls 20000 +trainer/Policy Loss -18.7465 +trainer/Log Pis Mean 23.834 +trainer/Log Pis Std 13.5527 +trainer/Log Pis Max 63.6481 +trainer/Log Pis Min -8.60583 +trainer/policy/mean Mean -0.0377699 +trainer/policy/mean Std 0.89436 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 3.00543 +trainer/policy/normal/std Std 0.460135 +trainer/policy/normal/std Max 5.22158 +trainer/policy/normal/std Min 0.837821 +trainer/policy/normal/log_std Mean 1.08626 +trainer/policy/normal/log_std Std 0.178885 +trainer/policy/normal/log_std Max 1.6528 +trainer/policy/normal/log_std Min -0.176951 +eval/num steps total 20000 +eval/num paths total 20 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.685467 +eval/Actions Std 0.674928 +eval/Actions Max 0.999978 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69877 +time/logging (s) 0.00366808 +time/sampling batch (s) 0.284353 +time/saving (s) 0.0034624 +time/training (s) 6.69204 +time/epoch (s) 9.68229 +time/total (s) 245.978 +Epoch -981 +---------------------------------- --------------- +2022-05-10 13:14:49.103128 PDT | [0] Epoch -980 finished +---------------------------------- --------------- +epoch -980 +replay_buffer/size 999033 +trainer/num train calls 21000 +trainer/Policy Loss -18.6884 +trainer/Log Pis Mean 24.8122 +trainer/Log Pis Std 13.6536 +trainer/Log Pis Max 70.846 +trainer/Log Pis Min -9.22607 +trainer/policy/mean Mean -0.0321888 +trainer/policy/mean Std 0.898715 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 3.04914 +trainer/policy/normal/std Std 0.485946 +trainer/policy/normal/std Max 5.29326 +trainer/policy/normal/std Min 0.847731 +trainer/policy/normal/log_std Mean 1.09923 +trainer/policy/normal/log_std Std 0.188913 +trainer/policy/normal/log_std Max 1.66643 +trainer/policy/normal/log_std Min -0.165192 +eval/num steps total 21000 +eval/num paths total 21 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0510304 +eval/Actions Std 0.929007 +eval/Actions Max 0.999999 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.27478 +time/logging (s) 0.00370971 +time/sampling batch (s) 0.284997 +time/saving (s) 0.00348517 +time/training (s) 6.81757 +time/epoch (s) 9.38455 +time/total (s) 255.365 +Epoch -980 +---------------------------------- --------------- +2022-05-10 13:14:59.529769 PDT | [0] Epoch -979 finished +---------------------------------- --------------- +epoch -979 +replay_buffer/size 999033 +trainer/num train calls 22000 +trainer/Policy Loss -19.2344 +trainer/Log Pis Mean 23.9161 +trainer/Log Pis Std 12.6328 +trainer/Log Pis Max 59.411 +trainer/Log Pis Min -5.55654 +trainer/policy/mean Mean -0.0384781 +trainer/policy/mean Std 0.898627 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 3.01159 +trainer/policy/normal/std Std 0.477904 +trainer/policy/normal/std Max 4.8014 +trainer/policy/normal/std Min 0.866252 +trainer/policy/normal/log_std Mean 1.08719 +trainer/policy/normal/log_std Std 0.185885 +trainer/policy/normal/log_std Max 1.56891 +trainer/policy/normal/log_std Min -0.14358 +eval/num steps total 22000 +eval/num paths total 22 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115483 +eval/Actions Std 0.901992 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52484 +time/logging (s) 0.00426193 +time/sampling batch (s) 0.284359 +time/saving (s) 0.00439009 +time/training (s) 7.59037 +time/epoch (s) 10.4082 +time/total (s) 265.776 +Epoch -979 +---------------------------------- --------------- +2022-05-10 13:15:10.016397 PDT | [0] Epoch -978 finished +---------------------------------- --------------- +epoch -978 +replay_buffer/size 999033 +trainer/num train calls 23000 +trainer/Policy Loss -18.8956 +trainer/Log Pis Mean 24.3136 +trainer/Log Pis Std 12.7878 +trainer/Log Pis Max 74.4968 +trainer/Log Pis Min -8.36527 +trainer/policy/mean Mean -0.0470551 +trainer/policy/mean Std 0.896061 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.99014 +trainer/policy/normal/std Std 0.481384 +trainer/policy/normal/std Max 5.02438 +trainer/policy/normal/std Min 0.772424 +trainer/policy/normal/log_std Mean 1.07885 +trainer/policy/normal/log_std Std 0.195838 +trainer/policy/normal/log_std Max 1.6143 +trainer/policy/normal/log_std Min -0.258222 +eval/num steps total 23000 +eval/num paths total 23 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167897 +eval/Actions Std 0.897788 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40172 +time/logging (s) 0.00377546 +time/sampling batch (s) 0.5347 +time/saving (s) 0.00368774 +time/training (s) 7.52321 +time/epoch (s) 10.4671 +time/total (s) 276.247 +Epoch -978 +---------------------------------- --------------- +2022-05-10 13:15:19.286524 PDT | [0] Epoch -977 finished +---------------------------------- --------------- +epoch -977 +replay_buffer/size 999033 +trainer/num train calls 24000 +trainer/Policy Loss -18.0327 +trainer/Log Pis Mean 25.0176 +trainer/Log Pis Std 13.582 +trainer/Log Pis Max 66.6791 +trainer/Log Pis Min -6.31524 +trainer/policy/mean Mean -0.0371484 +trainer/policy/mean Std 0.898269 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 3.03428 +trainer/policy/normal/std Std 0.466906 +trainer/policy/normal/std Max 4.96418 +trainer/policy/normal/std Min 0.811779 +trainer/policy/normal/log_std Mean 1.09533 +trainer/policy/normal/log_std Std 0.183202 +trainer/policy/normal/log_std Max 1.60225 +trainer/policy/normal/log_std Min -0.208527 +eval/num steps total 24000 +eval/num paths total 24 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0441096 +eval/Actions Std 0.883914 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7638 +time/logging (s) 0.00365325 +time/sampling batch (s) 0.281588 +time/saving (s) 0.00357427 +time/training (s) 6.19886 +time/epoch (s) 9.25148 +time/total (s) 285.501 +Epoch -977 +---------------------------------- --------------- +2022-05-10 13:15:29.088862 PDT | [0] Epoch -976 finished +---------------------------------- --------------- +epoch -976 +replay_buffer/size 999033 +trainer/num train calls 25000 +trainer/Policy Loss -19.6076 +trainer/Log Pis Mean 23.2235 +trainer/Log Pis Std 13.472 +trainer/Log Pis Max 68.6527 +trainer/Log Pis Min -7.10271 +trainer/policy/mean Mean -0.0509667 +trainer/policy/mean Std 0.897439 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.94037 +trainer/policy/normal/std Std 0.477158 +trainer/policy/normal/std Max 4.4006 +trainer/policy/normal/std Min 0.706433 +trainer/policy/normal/log_std Mean 1.06142 +trainer/policy/normal/log_std Std 0.20075 +trainer/policy/normal/log_std Max 1.48174 +trainer/policy/normal/log_std Min -0.347527 +eval/num steps total 25000 +eval/num paths total 25 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.466731 +eval/Actions Std 0.834611 +eval/Actions Max 0.999989 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58839 +time/logging (s) 0.00383085 +time/sampling batch (s) 0.282437 +time/saving (s) 0.00342053 +time/training (s) 6.90593 +time/epoch (s) 9.78401 +time/total (s) 295.288 +Epoch -976 +---------------------------------- --------------- +2022-05-10 13:15:38.921933 PDT | [0] Epoch -975 finished +---------------------------------- --------------- +epoch -975 +replay_buffer/size 999033 +trainer/num train calls 26000 +trainer/Policy Loss -18.8163 +trainer/Log Pis Mean 24.1421 +trainer/Log Pis Std 13.125 +trainer/Log Pis Max 63.9417 +trainer/Log Pis Min -6.99081 +trainer/policy/mean Mean -0.0289781 +trainer/policy/mean Std 0.897146 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.9748 +trainer/policy/normal/std Std 0.499805 +trainer/policy/normal/std Max 5.01031 +trainer/policy/normal/std Min 0.774942 +trainer/policy/normal/log_std Mean 1.0722 +trainer/policy/normal/log_std Std 0.20454 +trainer/policy/normal/log_std Max 1.6115 +trainer/policy/normal/log_std Min -0.254966 +eval/num steps total 26000 +eval/num paths total 26 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0592156 +eval/Actions Std 0.902776 +eval/Actions Max 0.999992 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53097 +time/logging (s) 0.00415824 +time/sampling batch (s) 0.282521 +time/saving (s) 0.00405037 +time/training (s) 6.99309 +time/epoch (s) 9.81478 +time/total (s) 305.105 +Epoch -975 +---------------------------------- --------------- +2022-05-10 13:15:49.552001 PDT | [0] Epoch -974 finished +---------------------------------- --------------- +epoch -974 +replay_buffer/size 999033 +trainer/num train calls 27000 +trainer/Policy Loss -19.3754 +trainer/Log Pis Mean 24.5047 +trainer/Log Pis Std 13.3475 +trainer/Log Pis Max 74.2306 +trainer/Log Pis Min -6.75855 +trainer/policy/mean Mean -0.0519948 +trainer/policy/mean Std 0.89354 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 3.04927 +trainer/policy/normal/std Std 0.507381 +trainer/policy/normal/std Max 7.09645 +trainer/policy/normal/std Min 0.814568 +trainer/policy/normal/log_std Mean 1.09816 +trainer/policy/normal/log_std Std 0.195183 +trainer/policy/normal/log_std Max 1.95959 +trainer/policy/normal/log_std Min -0.205097 +eval/num steps total 27000 +eval/num paths total 27 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.435272 +eval/Actions Std 0.880497 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60573 +time/logging (s) 0.00382952 +time/sampling batch (s) 0.283908 +time/saving (s) 0.00351385 +time/training (s) 7.71392 +time/epoch (s) 10.6109 +time/total (s) 315.719 +Epoch -974 +---------------------------------- --------------- +2022-05-10 13:16:00.393832 PDT | [0] Epoch -973 finished +---------------------------------- --------------- +epoch -973 +replay_buffer/size 999033 +trainer/num train calls 28000 +trainer/Policy Loss -19.34 +trainer/Log Pis Mean 23.396 +trainer/Log Pis Std 12.8646 +trainer/Log Pis Max 65.0776 +trainer/Log Pis Min -6.36592 +trainer/policy/mean Mean -0.0414306 +trainer/policy/mean Std 0.89952 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.93171 +trainer/policy/normal/std Std 0.477092 +trainer/policy/normal/std Max 4.62612 +trainer/policy/normal/std Min 0.684591 +trainer/policy/normal/log_std Mean 1.0585 +trainer/policy/normal/log_std Std 0.200601 +trainer/policy/normal/log_std Max 1.53172 +trainer/policy/normal/log_std Min -0.378933 +eval/num steps total 28000 +eval/num paths total 28 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0540757 +eval/Actions Std 0.910777 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63041 +time/logging (s) 0.00370257 +time/sampling batch (s) 0.290316 +time/saving (s) 0.00358118 +time/training (s) 7.89477 +time/epoch (s) 10.8228 +time/total (s) 326.545 +Epoch -973 +---------------------------------- --------------- +2022-05-10 13:16:12.125037 PDT | [0] Epoch -972 finished +---------------------------------- --------------- +epoch -972 +replay_buffer/size 999033 +trainer/num train calls 29000 +trainer/Policy Loss -19.2655 +trainer/Log Pis Mean 23.4411 +trainer/Log Pis Std 13.4503 +trainer/Log Pis Max 67.996 +trainer/Log Pis Min -9.42021 +trainer/policy/mean Mean -0.0152926 +trainer/policy/mean Std 0.899763 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.98688 +trainer/policy/normal/std Std 0.478306 +trainer/policy/normal/std Max 4.77854 +trainer/policy/normal/std Min 0.73604 +trainer/policy/normal/log_std Mean 1.07792 +trainer/policy/normal/log_std Std 0.194828 +trainer/policy/normal/log_std Max 1.56413 +trainer/policy/normal/log_std Min -0.30647 +eval/num steps total 29000 +eval/num paths total 29 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0783312 +eval/Actions Std 0.918828 +eval/Actions Max 0.999993 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73307 +time/logging (s) 0.00436987 +time/sampling batch (s) 0.781752 +time/saving (s) 0.00456298 +time/training (s) 8.18943 +time/epoch (s) 11.7132 +time/total (s) 338.261 +Epoch -972 +---------------------------------- --------------- +2022-05-10 13:16:22.324759 PDT | [0] Epoch -971 finished +---------------------------------- --------------- +epoch -971 +replay_buffer/size 999033 +trainer/num train calls 30000 +trainer/Policy Loss -19.3195 +trainer/Log Pis Mean 23.3555 +trainer/Log Pis Std 13.16 +trainer/Log Pis Max 64.4543 +trainer/Log Pis Min -9.26863 +trainer/policy/mean Mean -0.0472648 +trainer/policy/mean Std 0.897867 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.95653 +trainer/policy/normal/std Std 0.495645 +trainer/policy/normal/std Max 4.86826 +trainer/policy/normal/std Min 0.669273 +trainer/policy/normal/log_std Mean 1.06571 +trainer/policy/normal/log_std Std 0.207707 +trainer/policy/normal/log_std Max 1.58274 +trainer/policy/normal/log_std Min -0.401563 +eval/num steps total 30000 +eval/num paths total 30 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.278558 +eval/Actions Std 0.731927 +eval/Actions Max 0.999975 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68521 +time/logging (s) 0.00371858 +time/sampling batch (s) 0.529927 +time/saving (s) 0.00345384 +time/training (s) 6.95802 +time/epoch (s) 10.1803 +time/total (s) 348.445 +Epoch -971 +---------------------------------- --------------- +2022-05-10 13:16:32.579734 PDT | [0] Epoch -970 finished +---------------------------------- --------------- +epoch -970 +replay_buffer/size 999033 +trainer/num train calls 31000 +trainer/Policy Loss -20.5289 +trainer/Log Pis Mean 24.0539 +trainer/Log Pis Std 13.1305 +trainer/Log Pis Max 70.3689 +trainer/Log Pis Min -4.26228 +trainer/policy/mean Mean -0.0138022 +trainer/policy/mean Std 0.897477 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.95905 +trainer/policy/normal/std Std 0.501845 +trainer/policy/normal/std Max 6.74727 +trainer/policy/normal/std Min 0.734777 +trainer/policy/normal/log_std Mean 1.06738 +trainer/policy/normal/log_std Std 0.199659 +trainer/policy/normal/log_std Max 1.90914 +trainer/policy/normal/log_std Min -0.308188 +eval/num steps total 31000 +eval/num paths total 31 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0198916 +eval/Actions Std 0.888923 +eval/Actions Max 0.999997 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38177 +time/logging (s) 0.00373963 +time/sampling batch (s) 0.279311 +time/saving (s) 0.00346666 +time/training (s) 7.56821 +time/epoch (s) 10.2365 +time/total (s) 358.684 +Epoch -970 +---------------------------------- --------------- +2022-05-10 13:16:42.303363 PDT | [0] Epoch -969 finished +---------------------------------- --------------- +epoch -969 +replay_buffer/size 999033 +trainer/num train calls 32000 +trainer/Policy Loss -19.4424 +trainer/Log Pis Mean 24.5369 +trainer/Log Pis Std 13.4676 +trainer/Log Pis Max 69.9704 +trainer/Log Pis Min -11.0627 +trainer/policy/mean Mean -0.0597983 +trainer/policy/mean Std 0.900707 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.96002 +trainer/policy/normal/std Std 0.524831 +trainer/policy/normal/std Max 4.86812 +trainer/policy/normal/std Min 0.687505 +trainer/policy/normal/log_std Mean 1.06477 +trainer/policy/normal/log_std Std 0.219381 +trainer/policy/normal/log_std Max 1.58271 +trainer/policy/normal/log_std Min -0.374686 +eval/num steps total 32000 +eval/num paths total 32 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.10343 +eval/Actions Std 0.886207 +eval/Actions Max 0.999994 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67499 +time/logging (s) 0.00366568 +time/sampling batch (s) 0.281821 +time/saving (s) 0.00343021 +time/training (s) 6.74101 +time/epoch (s) 9.70491 +time/total (s) 368.392 +Epoch -969 +---------------------------------- --------------- +2022-05-10 13:16:52.119789 PDT | [0] Epoch -968 finished +---------------------------------- --------------- +epoch -968 +replay_buffer/size 999033 +trainer/num train calls 33000 +trainer/Policy Loss -20.0527 +trainer/Log Pis Mean 24.3296 +trainer/Log Pis Std 13.3133 +trainer/Log Pis Max 65.4578 +trainer/Log Pis Min -6.76592 +trainer/policy/mean Mean -0.0300477 +trainer/policy/mean Std 0.903414 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.96492 +trainer/policy/normal/std Std 0.504151 +trainer/policy/normal/std Max 6.37621 +trainer/policy/normal/std Min 0.712583 +trainer/policy/normal/log_std Mean 1.06901 +trainer/policy/normal/log_std Std 0.202412 +trainer/policy/normal/log_std Max 1.85257 +trainer/policy/normal/log_std Min -0.338859 +eval/num steps total 33000 +eval/num paths total 33 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0714706 +eval/Actions Std 0.887356 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6079 +time/logging (s) 0.00441023 +time/sampling batch (s) 0.284262 +time/saving (s) 0.00451155 +time/training (s) 6.89726 +time/epoch (s) 9.79834 +time/total (s) 378.193 +Epoch -968 +---------------------------------- --------------- +2022-05-10 13:17:02.321758 PDT | [0] Epoch -967 finished +---------------------------------- --------------- +epoch -967 +replay_buffer/size 999033 +trainer/num train calls 34000 +trainer/Policy Loss -18.6049 +trainer/Log Pis Mean 23.7748 +trainer/Log Pis Std 13.0213 +trainer/Log Pis Max 63.3471 +trainer/Log Pis Min -6.43176 +trainer/policy/mean Mean -0.0164484 +trainer/policy/mean Std 0.900702 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.94047 +trainer/policy/normal/std Std 0.503903 +trainer/policy/normal/std Max 5.32081 +trainer/policy/normal/std Min 0.641226 +trainer/policy/normal/log_std Mean 1.05943 +trainer/policy/normal/log_std Std 0.212562 +trainer/policy/normal/log_std Max 1.67163 +trainer/policy/normal/log_std Min -0.444374 +eval/num steps total 34000 +eval/num paths total 34 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.086323 +eval/Actions Std 0.898874 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66601 +time/logging (s) 0.00371962 +time/sampling batch (s) 0.280954 +time/saving (s) 0.00348069 +time/training (s) 7.22801 +time/epoch (s) 10.1822 +time/total (s) 388.379 +Epoch -967 +---------------------------------- --------------- +2022-05-10 13:17:12.320204 PDT | [0] Epoch -966 finished +---------------------------------- --------------- +epoch -966 +replay_buffer/size 999033 +trainer/num train calls 35000 +trainer/Policy Loss -18.9632 +trainer/Log Pis Mean 24.2028 +trainer/Log Pis Std 12.5714 +trainer/Log Pis Max 65.6 +trainer/Log Pis Min -6.4892 +trainer/policy/mean Mean -0.0331549 +trainer/policy/mean Std 0.899081 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.9407 +trainer/policy/normal/std Std 0.495438 +trainer/policy/normal/std Max 4.72955 +trainer/policy/normal/std Min 0.653791 +trainer/policy/normal/log_std Mean 1.06055 +trainer/policy/normal/log_std Std 0.205562 +trainer/policy/normal/log_std Max 1.55383 +trainer/policy/normal/log_std Min -0.424968 +eval/num steps total 35000 +eval/num paths total 35 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.210903 +eval/Actions Std 0.941512 +eval/Actions Max 0.999986 +eval/Actions Min -0.999911 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76716 +time/logging (s) 0.00419601 +time/sampling batch (s) 0.781586 +time/saving (s) 0.00408045 +time/training (s) 6.42326 +time/epoch (s) 9.98028 +time/total (s) 398.362 +Epoch -966 +---------------------------------- --------------- +2022-05-10 13:17:23.283305 PDT | [0] Epoch -965 finished +---------------------------------- -------------- +epoch -965 +replay_buffer/size 999033 +trainer/num train calls 36000 +trainer/Policy Loss -19.7967 +trainer/Log Pis Mean 24.2488 +trainer/Log Pis Std 12.9161 +trainer/Log Pis Max 62.8575 +trainer/Log Pis Min -8.26136 +trainer/policy/mean Mean -0.0408124 +trainer/policy/mean Std 0.902673 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.9169 +trainer/policy/normal/std Std 0.473609 +trainer/policy/normal/std Max 4.67404 +trainer/policy/normal/std Min 0.66317 +trainer/policy/normal/log_std Mean 1.05353 +trainer/policy/normal/log_std Std 0.199861 +trainer/policy/normal/log_std Max 1.54202 +trainer/policy/normal/log_std Min -0.410724 +eval/num steps total 36000 +eval/num paths total 36 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.080013 +eval/Actions Std 0.905563 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51454 +time/logging (s) 0.0048075 +time/sampling batch (s) 0.78568 +time/saving (s) 0.0038571 +time/training (s) 7.63546 +time/epoch (s) 10.9443 +time/total (s) 409.309 +Epoch -965 +---------------------------------- -------------- +2022-05-10 13:17:33.436827 PDT | [0] Epoch -964 finished +---------------------------------- --------------- +epoch -964 +replay_buffer/size 999033 +trainer/num train calls 37000 +trainer/Policy Loss -19.9436 +trainer/Log Pis Mean 24.8213 +trainer/Log Pis Std 13.9961 +trainer/Log Pis Max 71.0273 +trainer/Log Pis Min -6.97267 +trainer/policy/mean Mean -0.0253241 +trainer/policy/mean Std 0.897547 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.92359 +trainer/policy/normal/std Std 0.49659 +trainer/policy/normal/std Max 5.32633 +trainer/policy/normal/std Min 0.63715 +trainer/policy/normal/log_std Mean 1.05422 +trainer/policy/normal/log_std Std 0.209236 +trainer/policy/normal/log_std Max 1.67266 +trainer/policy/normal/log_std Min -0.450751 +eval/num steps total 37000 +eval/num paths total 37 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00509764 +eval/Actions Std 0.888204 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64276 +time/logging (s) 0.00379716 +time/sampling batch (s) 0.283229 +time/saving (s) 0.00358083 +time/training (s) 7.20017 +time/epoch (s) 10.1335 +time/total (s) 419.446 +Epoch -964 +---------------------------------- --------------- +2022-05-10 13:17:43.951909 PDT | [0] Epoch -963 finished +---------------------------------- --------------- +epoch -963 +replay_buffer/size 999033 +trainer/num train calls 38000 +trainer/Policy Loss -19.2987 +trainer/Log Pis Mean 24.1484 +trainer/Log Pis Std 13.2226 +trainer/Log Pis Max 65.0465 +trainer/Log Pis Min -6.17711 +trainer/policy/mean Mean -0.0260707 +trainer/policy/mean Std 0.903695 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.88657 +trainer/policy/normal/std Std 0.520432 +trainer/policy/normal/std Max 5.59544 +trainer/policy/normal/std Min 0.648126 +trainer/policy/normal/log_std Mean 1.03892 +trainer/policy/normal/log_std Std 0.223291 +trainer/policy/normal/log_std Max 1.72195 +trainer/policy/normal/log_std Min -0.43367 +eval/num steps total 38000 +eval/num paths total 38 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0262704 +eval/Actions Std 0.894838 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74495 +time/logging (s) 0.00410731 +time/sampling batch (s) 0.53516 +time/saving (s) 0.00412519 +time/training (s) 7.20805 +time/epoch (s) 10.4964 +time/total (s) 429.945 +Epoch -963 +---------------------------------- --------------- +2022-05-10 13:17:53.669615 PDT | [0] Epoch -962 finished +---------------------------------- --------------- +epoch -962 +replay_buffer/size 999033 +trainer/num train calls 39000 +trainer/Policy Loss -18.8587 +trainer/Log Pis Mean 23.8358 +trainer/Log Pis Std 13.2926 +trainer/Log Pis Max 66.8075 +trainer/Log Pis Min -5.79574 +trainer/policy/mean Mean -0.0302397 +trainer/policy/mean Std 0.897307 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90558 +trainer/policy/normal/std Std 0.489352 +trainer/policy/normal/std Max 4.95663 +trainer/policy/normal/std Min 0.608044 +trainer/policy/normal/log_std Mean 1.04817 +trainer/policy/normal/log_std Std 0.208978 +trainer/policy/normal/log_std Max 1.60073 +trainer/policy/normal/log_std Min -0.497508 +eval/num steps total 39000 +eval/num paths total 39 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101008 +eval/Actions Std 0.911672 +eval/Actions Max 0.999984 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87389 +time/logging (s) 0.00376745 +time/sampling batch (s) 0.288178 +time/saving (s) 0.00367587 +time/training (s) 6.52824 +time/epoch (s) 9.69775 +time/total (s) 439.646 +Epoch -962 +---------------------------------- --------------- +2022-05-10 13:18:04.243338 PDT | [0] Epoch -961 finished +---------------------------------- --------------- +epoch -961 +replay_buffer/size 999033 +trainer/num train calls 40000 +trainer/Policy Loss -19.0764 +trainer/Log Pis Mean 24.6891 +trainer/Log Pis Std 13.9427 +trainer/Log Pis Max 68.7798 +trainer/Log Pis Min -9.86785 +trainer/policy/mean Mean -0.0241281 +trainer/policy/mean Std 0.902127 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.95817 +trainer/policy/normal/std Std 0.512304 +trainer/policy/normal/std Max 4.79997 +trainer/policy/normal/std Min 0.60312 +trainer/policy/normal/log_std Mean 1.06537 +trainer/policy/normal/log_std Std 0.212418 +trainer/policy/normal/log_std Max 1.56861 +trainer/policy/normal/log_std Min -0.505639 +eval/num steps total 40000 +eval/num paths total 40 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00443868 +eval/Actions Std 0.885761 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65596 +time/logging (s) 0.00377355 +time/sampling batch (s) 0.286167 +time/saving (s) 0.00358324 +time/training (s) 7.60528 +time/epoch (s) 10.5548 +time/total (s) 450.204 +Epoch -961 +---------------------------------- --------------- +2022-05-10 13:18:13.733918 PDT | [0] Epoch -960 finished +---------------------------------- --------------- +epoch -960 +replay_buffer/size 999033 +trainer/num train calls 41000 +trainer/Policy Loss -19.5874 +trainer/Log Pis Mean 24.6839 +trainer/Log Pis Std 13.0393 +trainer/Log Pis Max 70.2738 +trainer/Log Pis Min -11.104 +trainer/policy/mean Mean -0.0396522 +trainer/policy/mean Std 0.903581 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.93685 +trainer/policy/normal/std Std 0.515147 +trainer/policy/normal/std Max 5.27807 +trainer/policy/normal/std Min 0.586294 +trainer/policy/normal/log_std Mean 1.05713 +trainer/policy/normal/log_std Std 0.219581 +trainer/policy/normal/log_std Max 1.66356 +trainer/policy/normal/log_std Min -0.533934 +eval/num steps total 41000 +eval/num paths total 41 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105368 +eval/Actions Std 0.893677 +eval/Actions Max 0.999987 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69106 +time/logging (s) 0.00428618 +time/sampling batch (s) 0.278219 +time/saving (s) 0.00428645 +time/training (s) 6.49474 +time/epoch (s) 9.47259 +time/total (s) 459.679 +Epoch -960 +---------------------------------- --------------- +2022-05-10 13:18:24.140630 PDT | [0] Epoch -959 finished +---------------------------------- --------------- +epoch -959 +replay_buffer/size 999033 +trainer/num train calls 42000 +trainer/Policy Loss -18.9392 +trainer/Log Pis Mean 24.564 +trainer/Log Pis Std 12.8547 +trainer/Log Pis Max 62.6789 +trainer/Log Pis Min -7.73131 +trainer/policy/mean Mean -0.0178451 +trainer/policy/mean Std 0.898254 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.97189 +trainer/policy/normal/std Std 0.514314 +trainer/policy/normal/std Max 5.31857 +trainer/policy/normal/std Min 0.577201 +trainer/policy/normal/log_std Mean 1.06972 +trainer/policy/normal/log_std Std 0.215074 +trainer/policy/normal/log_std Max 1.6712 +trainer/policy/normal/log_std Min -0.549564 +eval/num steps total 42000 +eval/num paths total 42 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.469074 +eval/Actions Std 0.782471 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64316 +time/logging (s) 0.00376457 +time/sampling batch (s) 0.277392 +time/saving (s) 0.0035815 +time/training (s) 7.45924 +time/epoch (s) 10.3871 +time/total (s) 470.07 +Epoch -959 +---------------------------------- --------------- +2022-05-10 13:18:34.381118 PDT | [0] Epoch -958 finished +---------------------------------- --------------- +epoch -958 +replay_buffer/size 999033 +trainer/num train calls 43000 +trainer/Policy Loss -18.7409 +trainer/Log Pis Mean 25.3874 +trainer/Log Pis Std 13.1057 +trainer/Log Pis Max 64.8144 +trainer/Log Pis Min -12.2656 +trainer/policy/mean Mean -0.0326841 +trainer/policy/mean Std 0.902204 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.96949 +trainer/policy/normal/std Std 0.505225 +trainer/policy/normal/std Max 4.80393 +trainer/policy/normal/std Min 0.563902 +trainer/policy/normal/log_std Mean 1.06912 +trainer/policy/normal/log_std Std 0.21493 +trainer/policy/normal/log_std Max 1.56943 +trainer/policy/normal/log_std Min -0.572875 +eval/num steps total 43000 +eval/num paths total 43 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.134735 +eval/Actions Std 0.898975 +eval/Actions Max 0.999965 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60351 +time/logging (s) 0.00376745 +time/sampling batch (s) 0.284316 +time/saving (s) 0.00347372 +time/training (s) 7.3266 +time/epoch (s) 10.2217 +time/total (s) 480.294 +Epoch -958 +---------------------------------- --------------- +2022-05-10 13:18:44.110548 PDT | [0] Epoch -957 finished +---------------------------------- --------------- +epoch -957 +replay_buffer/size 999033 +trainer/num train calls 44000 +trainer/Policy Loss -19.0074 +trainer/Log Pis Mean 23.5214 +trainer/Log Pis Std 13.1928 +trainer/Log Pis Max 69.4187 +trainer/Log Pis Min -7.66778 +trainer/policy/mean Mean -0.0503337 +trainer/policy/mean Std 0.901724 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85947 +trainer/policy/normal/std Std 0.483233 +trainer/policy/normal/std Max 5.02747 +trainer/policy/normal/std Min 0.55867 +trainer/policy/normal/log_std Mean 1.03229 +trainer/policy/normal/log_std Std 0.20771 +trainer/policy/normal/log_std Max 1.61492 +trainer/policy/normal/log_std Min -0.582197 +eval/num steps total 44000 +eval/num paths total 44 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122873 +eval/Actions Std 0.887642 +eval/Actions Max 0.999996 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4825 +time/logging (s) 0.00415923 +time/sampling batch (s) 0.280735 +time/saving (s) 0.00423559 +time/training (s) 6.93952 +time/epoch (s) 9.71115 +time/total (s) 490.008 +Epoch -957 +---------------------------------- --------------- +2022-05-10 13:18:54.085255 PDT | [0] Epoch -956 finished +---------------------------------- --------------- +epoch -956 +replay_buffer/size 999033 +trainer/num train calls 45000 +trainer/Policy Loss -19.0718 +trainer/Log Pis Mean 24.0123 +trainer/Log Pis Std 13.3196 +trainer/Log Pis Max 67.017 +trainer/Log Pis Min -11.2804 +trainer/policy/mean Mean -0.0302052 +trainer/policy/mean Std 0.898866 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.95361 +trainer/policy/normal/std Std 0.541509 +trainer/policy/normal/std Max 5.14891 +trainer/policy/normal/std Min 0.579885 +trainer/policy/normal/log_std Mean 1.0611 +trainer/policy/normal/log_std Std 0.228089 +trainer/policy/normal/log_std Max 1.63878 +trainer/policy/normal/log_std Min -0.544926 +eval/num steps total 45000 +eval/num paths total 45 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.457544 +eval/Actions Std 0.877642 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81582 +time/logging (s) 0.00386561 +time/sampling batch (s) 0.533187 +time/saving (s) 0.00364068 +time/training (s) 6.5987 +time/epoch (s) 9.95521 +time/total (s) 499.967 +Epoch -956 +---------------------------------- --------------- +2022-05-10 13:19:04.209777 PDT | [0] Epoch -955 finished +---------------------------------- --------------- +epoch -955 +replay_buffer/size 999033 +trainer/num train calls 46000 +trainer/Policy Loss -19.0439 +trainer/Log Pis Mean 24.0113 +trainer/Log Pis Std 13.3836 +trainer/Log Pis Max 65.0824 +trainer/Log Pis Min -7.99655 +trainer/policy/mean Mean -0.0545122 +trainer/policy/mean Std 0.896832 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.97415 +trainer/policy/normal/std Std 0.524181 +trainer/policy/normal/std Max 5.06789 +trainer/policy/normal/std Min 0.629939 +trainer/policy/normal/log_std Mean 1.06988 +trainer/policy/normal/log_std Std 0.217617 +trainer/policy/normal/log_std Max 1.62292 +trainer/policy/normal/log_std Min -0.462132 +eval/num steps total 46000 +eval/num paths total 46 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.122145 +eval/Actions Std 0.812816 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66413 +time/logging (s) 0.00371783 +time/sampling batch (s) 0.28287 +time/saving (s) 0.00345105 +time/training (s) 7.15155 +time/epoch (s) 10.1057 +time/total (s) 510.075 +Epoch -955 +---------------------------------- --------------- +2022-05-10 13:19:14.487149 PDT | [0] Epoch -954 finished +---------------------------------- --------------- +epoch -954 +replay_buffer/size 999033 +trainer/num train calls 47000 +trainer/Policy Loss -20.4502 +trainer/Log Pis Mean 25.0689 +trainer/Log Pis Std 13.4844 +trainer/Log Pis Max 65.0684 +trainer/Log Pis Min -8.18362 +trainer/policy/mean Mean -0.0508176 +trainer/policy/mean Std 0.903804 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.93314 +trainer/policy/normal/std Std 0.514219 +trainer/policy/normal/std Max 5.34869 +trainer/policy/normal/std Min 0.533665 +trainer/policy/normal/log_std Mean 1.05641 +trainer/policy/normal/log_std Std 0.214897 +trainer/policy/normal/log_std Max 1.67685 +trainer/policy/normal/log_std Min -0.627988 +eval/num steps total 47000 +eval/num paths total 47 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.255331 +eval/Actions Std 0.913665 +eval/Actions Max 0.999998 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77474 +time/logging (s) 0.00437512 +time/sampling batch (s) 0.281717 +time/saving (s) 0.00438244 +time/training (s) 7.19399 +time/epoch (s) 10.2592 +time/total (s) 520.337 +Epoch -954 +---------------------------------- --------------- +2022-05-10 13:19:24.102093 PDT | [0] Epoch -953 finished +---------------------------------- --------------- +epoch -953 +replay_buffer/size 999033 +trainer/num train calls 48000 +trainer/Policy Loss -19.2468 +trainer/Log Pis Mean 24.4319 +trainer/Log Pis Std 13.1951 +trainer/Log Pis Max 76.727 +trainer/Log Pis Min -9.74702 +trainer/policy/mean Mean -0.0059825 +trainer/policy/mean Std 0.899191 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.97457 +trainer/policy/normal/std Std 0.533475 +trainer/policy/normal/std Max 4.76745 +trainer/policy/normal/std Min 0.584989 +trainer/policy/normal/log_std Mean 1.06861 +trainer/policy/normal/log_std Std 0.227081 +trainer/policy/normal/log_std Max 1.56181 +trainer/policy/normal/log_std Min -0.536161 +eval/num steps total 48000 +eval/num paths total 48 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129713 +eval/Actions Std 0.87579 +eval/Actions Max 0.999986 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74773 +time/logging (s) 0.00383174 +time/sampling batch (s) 0.286222 +time/saving (s) 0.00362508 +time/training (s) 6.55331 +time/epoch (s) 9.59471 +time/total (s) 529.936 +Epoch -953 +---------------------------------- --------------- +2022-05-10 13:19:33.731658 PDT | [0] Epoch -952 finished +---------------------------------- --------------- +epoch -952 +replay_buffer/size 999033 +trainer/num train calls 49000 +trainer/Policy Loss -19.6007 +trainer/Log Pis Mean 23.476 +trainer/Log Pis Std 13.188 +trainer/Log Pis Max 70.1221 +trainer/Log Pis Min -8.47525 +trainer/policy/mean Mean -0.0350698 +trainer/policy/mean Std 0.901289 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84943 +trainer/policy/normal/std Std 0.491904 +trainer/policy/normal/std Max 4.98763 +trainer/policy/normal/std Min 0.588598 +trainer/policy/normal/log_std Mean 1.02793 +trainer/policy/normal/log_std Std 0.212714 +trainer/policy/normal/log_std Max 1.60696 +trainer/policy/normal/log_std Min -0.530012 +eval/num steps total 49000 +eval/num paths total 49 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.212781 +eval/Actions Std 0.838711 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56644 +time/logging (s) 0.00377584 +time/sampling batch (s) 0.286622 +time/saving (s) 0.00348727 +time/training (s) 6.75021 +time/epoch (s) 9.61053 +time/total (s) 539.549 +Epoch -952 +---------------------------------- --------------- +2022-05-10 13:19:44.585110 PDT | [0] Epoch -951 finished +---------------------------------- --------------- +epoch -951 +replay_buffer/size 999033 +trainer/num train calls 50000 +trainer/Policy Loss -18.0048 +trainer/Log Pis Mean 24.9866 +trainer/Log Pis Std 12.6864 +trainer/Log Pis Max 60.7602 +trainer/Log Pis Min -5.81781 +trainer/policy/mean Mean -0.0509965 +trainer/policy/mean Std 0.901905 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.92162 +trainer/policy/normal/std Std 0.522481 +trainer/policy/normal/std Max 4.60256 +trainer/policy/normal/std Min 0.599888 +trainer/policy/normal/log_std Mean 1.05108 +trainer/policy/normal/log_std Std 0.223451 +trainer/policy/normal/log_std Max 1.52661 +trainer/policy/normal/log_std Min -0.511012 +eval/num steps total 50000 +eval/num paths total 50 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.207137 +eval/Actions Std 0.889807 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.8332 +time/logging (s) 0.00377115 +time/sampling batch (s) 0.279777 +time/saving (s) 0.0034331 +time/training (s) 7.71462 +time/epoch (s) 10.8348 +time/total (s) 550.387 +Epoch -951 +---------------------------------- --------------- +2022-05-10 13:19:54.013051 PDT | [0] Epoch -950 finished +---------------------------------- --------------- +epoch -950 +replay_buffer/size 999033 +trainer/num train calls 51000 +trainer/Policy Loss -18.7766 +trainer/Log Pis Mean 23.9686 +trainer/Log Pis Std 13.4779 +trainer/Log Pis Max 63.2412 +trainer/Log Pis Min -7.75238 +trainer/policy/mean Mean -0.0344167 +trainer/policy/mean Std 0.900165 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.92369 +trainer/policy/normal/std Std 0.517131 +trainer/policy/normal/std Max 4.94145 +trainer/policy/normal/std Min 0.63716 +trainer/policy/normal/log_std Mean 1.05258 +trainer/policy/normal/log_std Std 0.218459 +trainer/policy/normal/log_std Max 1.59766 +trainer/policy/normal/log_std Min -0.450734 +eval/num steps total 51000 +eval/num paths total 51 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.199446 +eval/Actions Std 0.867492 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80053 +time/logging (s) 0.00410812 +time/sampling batch (s) 0.281472 +time/saving (s) 0.00428301 +time/training (s) 6.31915 +time/epoch (s) 9.40954 +time/total (s) 559.799 +Epoch -950 +---------------------------------- --------------- +2022-05-10 13:20:03.885622 PDT | [0] Epoch -949 finished +---------------------------------- --------------- +epoch -949 +replay_buffer/size 999033 +trainer/num train calls 52000 +trainer/Policy Loss -19.9826 +trainer/Log Pis Mean 24.192 +trainer/Log Pis Std 13.3975 +trainer/Log Pis Max 78.4668 +trainer/Log Pis Min -6.408 +trainer/policy/mean Mean -0.055701 +trainer/policy/mean Std 0.901744 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89359 +trainer/policy/normal/std Std 0.512181 +trainer/policy/normal/std Max 4.58516 +trainer/policy/normal/std Min 0.626964 +trainer/policy/normal/log_std Mean 1.04123 +trainer/policy/normal/log_std Std 0.22659 +trainer/policy/normal/log_std Max 1.52282 +trainer/policy/normal/log_std Min -0.466867 +eval/num steps total 52000 +eval/num paths total 52 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.435371 +eval/Actions Std 0.850935 +eval/Actions Max 0.999986 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6748 +time/logging (s) 0.00378632 +time/sampling batch (s) 0.536344 +time/saving (s) 0.0034296 +time/training (s) 6.63443 +time/epoch (s) 9.85279 +time/total (s) 569.655 +Epoch -949 +---------------------------------- --------------- +2022-05-10 13:20:14.324632 PDT | [0] Epoch -948 finished +---------------------------------- --------------- +epoch -948 +replay_buffer/size 999033 +trainer/num train calls 53000 +trainer/Policy Loss -19.9651 +trainer/Log Pis Mean 23.8824 +trainer/Log Pis Std 13.3483 +trainer/Log Pis Max 67.4937 +trainer/Log Pis Min -4.86276 +trainer/policy/mean Mean -0.0268228 +trainer/policy/mean Std 0.904584 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89069 +trainer/policy/normal/std Std 0.517114 +trainer/policy/normal/std Max 6.5836 +trainer/policy/normal/std Min 0.625544 +trainer/policy/normal/log_std Mean 1.04073 +trainer/policy/normal/log_std Std 0.221575 +trainer/policy/normal/log_std Max 1.88458 +trainer/policy/normal/log_std Min -0.469134 +eval/num steps total 53000 +eval/num paths total 53 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.348881 +eval/Actions Std 0.87392 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78058 +time/logging (s) 0.00375149 +time/sampling batch (s) 0.288413 +time/saving (s) 0.00349569 +time/training (s) 7.34337 +time/epoch (s) 10.4196 +time/total (s) 580.077 +Epoch -948 +---------------------------------- --------------- +2022-05-10 13:20:25.144649 PDT | [0] Epoch -947 finished +---------------------------------- --------------- +epoch -947 +replay_buffer/size 999033 +trainer/num train calls 54000 +trainer/Policy Loss -18.8356 +trainer/Log Pis Mean 23.8727 +trainer/Log Pis Std 12.9194 +trainer/Log Pis Max 71.3828 +trainer/Log Pis Min -6.59152 +trainer/policy/mean Mean -0.0330487 +trainer/policy/mean Std 0.898054 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.90699 +trainer/policy/normal/std Std 0.54741 +trainer/policy/normal/std Max 4.91309 +trainer/policy/normal/std Min 0.552624 +trainer/policy/normal/log_std Mean 1.04304 +trainer/policy/normal/log_std Std 0.242056 +trainer/policy/normal/log_std Max 1.5919 +trainer/policy/normal/log_std Min -0.593078 +eval/num steps total 54000 +eval/num paths total 54 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0952939 +eval/Actions Std 0.90156 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64568 +time/logging (s) 0.00415657 +time/sampling batch (s) 0.288586 +time/saving (s) 0.0042372 +time/training (s) 7.85853 +time/epoch (s) 10.8012 +time/total (s) 590.881 +Epoch -947 +---------------------------------- --------------- +2022-05-10 13:20:35.865902 PDT | [0] Epoch -946 finished +---------------------------------- ---------------- +epoch -946 +replay_buffer/size 999033 +trainer/num train calls 55000 +trainer/Policy Loss -18.2003 +trainer/Log Pis Mean 24.2003 +trainer/Log Pis Std 12.9964 +trainer/Log Pis Max 67.2758 +trainer/Log Pis Min -7.17675 +trainer/policy/mean Mean -0.0174919 +trainer/policy/mean Std 0.901801 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88037 +trainer/policy/normal/std Std 0.518683 +trainer/policy/normal/std Max 4.83021 +trainer/policy/normal/std Min 0.530008 +trainer/policy/normal/log_std Mean 1.03708 +trainer/policy/normal/log_std Std 0.221142 +trainer/policy/normal/log_std Max 1.57489 +trainer/policy/normal/log_std Min -0.634864 +eval/num steps total 55000 +eval/num paths total 55 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000373512 +eval/Actions Std 0.936558 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67993 +time/logging (s) 0.00380505 +time/sampling batch (s) 0.285825 +time/saving (s) 0.00346694 +time/training (s) 7.72848 +time/epoch (s) 10.7015 +time/total (s) 601.586 +Epoch -946 +---------------------------------- ---------------- +2022-05-10 13:20:46.499796 PDT | [0] Epoch -945 finished +---------------------------------- --------------- +epoch -945 +replay_buffer/size 999033 +trainer/num train calls 56000 +trainer/Policy Loss -18.9181 +trainer/Log Pis Mean 24.2339 +trainer/Log Pis Std 12.9906 +trainer/Log Pis Max 77.7495 +trainer/Log Pis Min -9.77387 +trainer/policy/mean Mean -0.0190146 +trainer/policy/mean Std 0.904249 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.92171 +trainer/policy/normal/std Std 0.533092 +trainer/policy/normal/std Max 6.20921 +trainer/policy/normal/std Min 0.547636 +trainer/policy/normal/log_std Mean 1.05017 +trainer/policy/normal/log_std Std 0.229254 +trainer/policy/normal/log_std Max 1.82603 +trainer/policy/normal/log_std Min -0.602145 +eval/num steps total 56000 +eval/num paths total 56 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.227416 +eval/Actions Std 0.92928 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68338 +time/logging (s) 0.00377993 +time/sampling batch (s) 0.532472 +time/saving (s) 0.00335071 +time/training (s) 7.39222 +time/epoch (s) 10.6152 +time/total (s) 612.204 +Epoch -945 +---------------------------------- --------------- +2022-05-10 13:20:56.359073 PDT | [0] Epoch -944 finished +---------------------------------- --------------- +epoch -944 +replay_buffer/size 999033 +trainer/num train calls 57000 +trainer/Policy Loss -19.25 +trainer/Log Pis Mean 24.0294 +trainer/Log Pis Std 13.1502 +trainer/Log Pis Max 69.8019 +trainer/Log Pis Min -7.3374 +trainer/policy/mean Mean -0.0605791 +trainer/policy/mean Std 0.901918 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.95682 +trainer/policy/normal/std Std 0.528452 +trainer/policy/normal/std Max 4.78498 +trainer/policy/normal/std Min 0.552349 +trainer/policy/normal/log_std Mean 1.06247 +trainer/policy/normal/log_std Std 0.22866 +trainer/policy/normal/log_std Max 1.56548 +trainer/policy/normal/log_std Min -0.593575 +eval/num steps total 57000 +eval/num paths total 57 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0173668 +eval/Actions Std 0.880909 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7107 +time/logging (s) 0.00403297 +time/sampling batch (s) 0.283896 +time/saving (s) 0.00428613 +time/training (s) 6.83782 +time/epoch (s) 9.84073 +time/total (s) 622.048 +Epoch -944 +---------------------------------- --------------- +2022-05-10 13:21:07.656795 PDT | [0] Epoch -943 finished +---------------------------------- --------------- +epoch -943 +replay_buffer/size 999033 +trainer/num train calls 58000 +trainer/Policy Loss -18.7019 +trainer/Log Pis Mean 23.8657 +trainer/Log Pis Std 12.7365 +trainer/Log Pis Max 64.2205 +trainer/Log Pis Min -9.95508 +trainer/policy/mean Mean -0.0243463 +trainer/policy/mean Std 0.898754 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.90481 +trainer/policy/normal/std Std 0.515089 +trainer/policy/normal/std Max 4.94491 +trainer/policy/normal/std Min 0.524082 +trainer/policy/normal/log_std Mean 1.04569 +trainer/policy/normal/log_std Std 0.222234 +trainer/policy/normal/log_std Max 1.59836 +trainer/policy/normal/log_std Min -0.646107 +eval/num steps total 58000 +eval/num paths total 58 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0962614 +eval/Actions Std 0.871856 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73224 +time/logging (s) 0.00368886 +time/sampling batch (s) 0.541712 +time/saving (s) 0.00354323 +time/training (s) 7.9965 +time/epoch (s) 11.2777 +time/total (s) 633.329 +Epoch -943 +---------------------------------- --------------- +2022-05-10 13:21:17.567898 PDT | [0] Epoch -942 finished +---------------------------------- --------------- +epoch -942 +replay_buffer/size 999033 +trainer/num train calls 59000 +trainer/Policy Loss -18.8966 +trainer/Log Pis Mean 24.7621 +trainer/Log Pis Std 12.866 +trainer/Log Pis Max 67.6377 +trainer/Log Pis Min -8.56504 +trainer/policy/mean Mean -0.0403923 +trainer/policy/mean Std 0.904189 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.86993 +trainer/policy/normal/std Std 0.53492 +trainer/policy/normal/std Max 4.80461 +trainer/policy/normal/std Min 0.558044 +trainer/policy/normal/log_std Mean 1.03117 +trainer/policy/normal/log_std Std 0.235285 +trainer/policy/normal/log_std Max 1.56958 +trainer/policy/normal/log_std Min -0.583318 +eval/num steps total 59000 +eval/num paths total 59 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0148374 +eval/Actions Std 0.919806 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4333 +time/logging (s) 0.00412552 +time/sampling batch (s) 0.284286 +time/saving (s) 0.00416178 +time/training (s) 7.16672 +time/epoch (s) 9.8926 +time/total (s) 643.224 +Epoch -942 +---------------------------------- --------------- +2022-05-10 13:21:26.857593 PDT | [0] Epoch -941 finished +---------------------------------- --------------- +epoch -941 +replay_buffer/size 999033 +trainer/num train calls 60000 +trainer/Policy Loss -18.4263 +trainer/Log Pis Mean 24.6266 +trainer/Log Pis Std 13.1605 +trainer/Log Pis Max 77.8612 +trainer/Log Pis Min -4.63763 +trainer/policy/mean Mean -0.0446673 +trainer/policy/mean Std 0.906477 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.89791 +trainer/policy/normal/std Std 0.510564 +trainer/policy/normal/std Max 4.7761 +trainer/policy/normal/std Min 0.509194 +trainer/policy/normal/log_std Mean 1.04302 +trainer/policy/normal/log_std Std 0.225416 +trainer/policy/normal/log_std Max 1.56362 +trainer/policy/normal/log_std Min -0.674927 +eval/num steps total 60000 +eval/num paths total 60 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.24722 +eval/Actions Std 0.934409 +eval/Actions Max 0.999949 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61215 +time/logging (s) 0.00377923 +time/sampling batch (s) 0.280993 +time/saving (s) 0.00354774 +time/training (s) 6.36976 +time/epoch (s) 9.27024 +time/total (s) 652.498 +Epoch -941 +---------------------------------- --------------- +2022-05-10 13:21:35.995786 PDT | [0] Epoch -940 finished +---------------------------------- --------------- +epoch -940 +replay_buffer/size 999033 +trainer/num train calls 61000 +trainer/Policy Loss -18.6187 +trainer/Log Pis Mean 24.2677 +trainer/Log Pis Std 12.8296 +trainer/Log Pis Max 65.4545 +trainer/Log Pis Min -6.96145 +trainer/policy/mean Mean -0.0441418 +trainer/policy/mean Std 0.906347 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.95958 +trainer/policy/normal/std Std 0.531235 +trainer/policy/normal/std Max 5.48437 +trainer/policy/normal/std Min 0.602854 +trainer/policy/normal/log_std Mean 1.06389 +trainer/policy/normal/log_std Std 0.224865 +trainer/policy/normal/log_std Max 1.7019 +trainer/policy/normal/log_std Min -0.50608 +eval/num steps total 61000 +eval/num paths total 61 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0511026 +eval/Actions Std 0.892186 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48479 +time/logging (s) 0.00380986 +time/sampling batch (s) 0.283636 +time/saving (s) 0.00354586 +time/training (s) 6.34375 +time/epoch (s) 9.11953 +time/total (s) 661.62 +Epoch -940 +---------------------------------- --------------- +2022-05-10 13:21:46.393933 PDT | [0] Epoch -939 finished +---------------------------------- --------------- +epoch -939 +replay_buffer/size 999033 +trainer/num train calls 62000 +trainer/Policy Loss -18.3916 +trainer/Log Pis Mean 24.1382 +trainer/Log Pis Std 12.7878 +trainer/Log Pis Max 64.6542 +trainer/Log Pis Min -6.93976 +trainer/policy/mean Mean -0.0164031 +trainer/policy/mean Std 0.899888 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8873 +trainer/policy/normal/std Std 0.513327 +trainer/policy/normal/std Max 5.88946 +trainer/policy/normal/std Min 0.627332 +trainer/policy/normal/log_std Mean 1.03959 +trainer/policy/normal/log_std Std 0.22167 +trainer/policy/normal/log_std Max 1.77316 +trainer/policy/normal/log_std Min -0.46628 +eval/num steps total 62000 +eval/num paths total 62 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167598 +eval/Actions Std 0.910537 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58698 +time/logging (s) 0.00437391 +time/sampling batch (s) 0.278883 +time/saving (s) 0.00429385 +time/training (s) 7.50559 +time/epoch (s) 10.3801 +time/total (s) 672.003 +Epoch -939 +---------------------------------- --------------- +2022-05-10 13:21:55.855652 PDT | [0] Epoch -938 finished +---------------------------------- --------------- +epoch -938 +replay_buffer/size 999033 +trainer/num train calls 63000 +trainer/Policy Loss -20.0205 +trainer/Log Pis Mean 24.4436 +trainer/Log Pis Std 13.9434 +trainer/Log Pis Max 78.1232 +trainer/Log Pis Min -10.6598 +trainer/policy/mean Mean -0.0405386 +trainer/policy/mean Std 0.906727 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.88738 +trainer/policy/normal/std Std 0.52129 +trainer/policy/normal/std Max 5.28071 +trainer/policy/normal/std Min 0.445149 +trainer/policy/normal/log_std Mean 1.03851 +trainer/policy/normal/log_std Std 0.229207 +trainer/policy/normal/log_std Max 1.66406 +trainer/policy/normal/log_std Min -0.809347 +eval/num steps total 63000 +eval/num paths total 63 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.46066 +eval/Actions Std 0.70129 +eval/Actions Max 0.999957 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53227 +time/logging (s) 0.00371567 +time/sampling batch (s) 0.282497 +time/saving (s) 0.00357883 +time/training (s) 6.61952 +time/epoch (s) 9.44158 +time/total (s) 681.448 +Epoch -938 +---------------------------------- --------------- +2022-05-10 13:22:04.658469 PDT | [0] Epoch -937 finished +---------------------------------- --------------- +epoch -937 +replay_buffer/size 999033 +trainer/num train calls 64000 +trainer/Policy Loss -19.4419 +trainer/Log Pis Mean 24.6287 +trainer/Log Pis Std 13.2241 +trainer/Log Pis Max 65.9507 +trainer/Log Pis Min -3.89562 +trainer/policy/mean Mean -0.0452357 +trainer/policy/mean Std 0.90429 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88471 +trainer/policy/normal/std Std 0.526879 +trainer/policy/normal/std Max 5.72071 +trainer/policy/normal/std Min 0.623256 +trainer/policy/normal/log_std Mean 1.03723 +trainer/policy/normal/log_std Std 0.230123 +trainer/policy/normal/log_std Max 1.74409 +trainer/policy/normal/log_std Min -0.472798 +eval/num steps total 64000 +eval/num paths total 64 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.313716 +eval/Actions Std 0.883704 +eval/Actions Max 0.999978 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46502 +time/logging (s) 0.00375288 +time/sampling batch (s) 0.284898 +time/saving (s) 0.00348491 +time/training (s) 6.02668 +time/epoch (s) 8.78383 +time/total (s) 690.235 +Epoch -937 +---------------------------------- --------------- +2022-05-10 13:22:15.134463 PDT | [0] Epoch -936 finished +---------------------------------- --------------- +epoch -936 +replay_buffer/size 999033 +trainer/num train calls 65000 +trainer/Policy Loss -18.6317 +trainer/Log Pis Mean 24.276 +trainer/Log Pis Std 12.5968 +trainer/Log Pis Max 63.95 +trainer/Log Pis Min -5.15273 +trainer/policy/mean Mean -0.05251 +trainer/policy/mean Std 0.898888 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87212 +trainer/policy/normal/std Std 0.526663 +trainer/policy/normal/std Max 5.05644 +trainer/policy/normal/std Min 0.520824 +trainer/policy/normal/log_std Mean 1.03272 +trainer/policy/normal/log_std Std 0.231227 +trainer/policy/normal/log_std Max 1.62066 +trainer/policy/normal/log_std Min -0.652344 +eval/num steps total 65000 +eval/num paths total 65 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125655 +eval/Actions Std 0.903735 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67707 +time/logging (s) 0.00398645 +time/sampling batch (s) 0.282786 +time/saving (s) 0.00349498 +time/training (s) 7.48997 +time/epoch (s) 10.4573 +time/total (s) 700.695 +Epoch -936 +---------------------------------- --------------- +2022-05-10 13:22:24.594525 PDT | [0] Epoch -935 finished +---------------------------------- --------------- +epoch -935 +replay_buffer/size 999033 +trainer/num train calls 66000 +trainer/Policy Loss -19.3855 +trainer/Log Pis Mean 24.2865 +trainer/Log Pis Std 14.0193 +trainer/Log Pis Max 67.7454 +trainer/Log Pis Min -12.5437 +trainer/policy/mean Mean -0.0584637 +trainer/policy/mean Std 0.903447 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.90543 +trainer/policy/normal/std Std 0.48755 +trainer/policy/normal/std Max 4.98741 +trainer/policy/normal/std Min 0.577081 +trainer/policy/normal/log_std Mean 1.04874 +trainer/policy/normal/log_std Std 0.203996 +trainer/policy/normal/log_std Max 1.60692 +trainer/policy/normal/log_std Min -0.549772 +eval/num steps total 66000 +eval/num paths total 66 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0967939 +eval/Actions Std 0.914094 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68083 +time/logging (s) 0.00426246 +time/sampling batch (s) 0.28328 +time/saving (s) 0.00438556 +time/training (s) 6.46855 +time/epoch (s) 9.44131 +time/total (s) 710.139 +Epoch -935 +---------------------------------- --------------- +2022-05-10 13:22:34.154825 PDT | [0] Epoch -934 finished +---------------------------------- --------------- +epoch -934 +replay_buffer/size 999033 +trainer/num train calls 67000 +trainer/Policy Loss -20.45 +trainer/Log Pis Mean 25.3808 +trainer/Log Pis Std 14.1385 +trainer/Log Pis Max 68.6904 +trainer/Log Pis Min -8.50606 +trainer/policy/mean Mean -0.0394305 +trainer/policy/mean Std 0.907985 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.97026 +trainer/policy/normal/std Std 0.530505 +trainer/policy/normal/std Max 5.19299 +trainer/policy/normal/std Min 0.487252 +trainer/policy/normal/log_std Mean 1.06716 +trainer/policy/normal/log_std Std 0.228285 +trainer/policy/normal/log_std Max 1.64731 +trainer/policy/normal/log_std Min -0.718974 +eval/num steps total 67000 +eval/num paths total 67 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.067059 +eval/Actions Std 0.905825 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75534 +time/logging (s) 0.00379539 +time/sampling batch (s) 0.28214 +time/saving (s) 0.00370572 +time/training (s) 6.49531 +time/epoch (s) 9.54029 +time/total (s) 719.683 +Epoch -934 +---------------------------------- --------------- +2022-05-10 13:22:44.069488 PDT | [0] Epoch -933 finished +---------------------------------- --------------- +epoch -933 +replay_buffer/size 999033 +trainer/num train calls 68000 +trainer/Policy Loss -18.7427 +trainer/Log Pis Mean 24.4211 +trainer/Log Pis Std 12.7331 +trainer/Log Pis Max 62.0667 +trainer/Log Pis Min -4.26871 +trainer/policy/mean Mean -0.0479817 +trainer/policy/mean Std 0.901287 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87241 +trainer/policy/normal/std Std 0.524684 +trainer/policy/normal/std Max 4.7947 +trainer/policy/normal/std Min 0.536205 +trainer/policy/normal/log_std Mean 1.03222 +trainer/policy/normal/log_std Std 0.236477 +trainer/policy/normal/log_std Max 1.56751 +trainer/policy/normal/log_std Min -0.623239 +eval/num steps total 68000 +eval/num paths total 68 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.44894 +eval/Actions Std 0.849701 +eval/Actions Max 0.999906 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67807 +time/logging (s) 0.00373002 +time/sampling batch (s) 0.28296 +time/saving (s) 0.00351431 +time/training (s) 6.9272 +time/epoch (s) 9.89548 +time/total (s) 729.581 +Epoch -933 +---------------------------------- --------------- +2022-05-10 13:22:54.475960 PDT | [0] Epoch -932 finished +---------------------------------- --------------- +epoch -932 +replay_buffer/size 999033 +trainer/num train calls 69000 +trainer/Policy Loss -19.4987 +trainer/Log Pis Mean 25.1842 +trainer/Log Pis Std 13.7413 +trainer/Log Pis Max 69.4465 +trainer/Log Pis Min -12.1293 +trainer/policy/mean Mean -0.0371469 +trainer/policy/mean Std 0.907385 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.95277 +trainer/policy/normal/std Std 0.5299 +trainer/policy/normal/std Max 4.98438 +trainer/policy/normal/std Min 0.587439 +trainer/policy/normal/log_std Mean 1.06166 +trainer/policy/normal/log_std Std 0.223897 +trainer/policy/normal/log_std Max 1.60631 +trainer/policy/normal/log_std Min -0.531983 +eval/num steps total 69000 +eval/num paths total 69 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0121223 +eval/Actions Std 0.914551 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66877 +time/logging (s) 0.00423053 +time/sampling batch (s) 0.533559 +time/saving (s) 0.00438111 +time/training (s) 7.17696 +time/epoch (s) 10.3879 +time/total (s) 739.972 +Epoch -932 +---------------------------------- --------------- +2022-05-10 13:23:04.712494 PDT | [0] Epoch -931 finished +---------------------------------- --------------- +epoch -931 +replay_buffer/size 999033 +trainer/num train calls 70000 +trainer/Policy Loss -18.4909 +trainer/Log Pis Mean 25.469 +trainer/Log Pis Std 13.4996 +trainer/Log Pis Max 65.366 +trainer/Log Pis Min -8.08009 +trainer/policy/mean Mean -0.0499603 +trainer/policy/mean Std 0.903933 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.9293 +trainer/policy/normal/std Std 0.538196 +trainer/policy/normal/std Max 6.10421 +trainer/policy/normal/std Min 0.448852 +trainer/policy/normal/log_std Mean 1.05249 +trainer/policy/normal/log_std Std 0.231364 +trainer/policy/normal/log_std Max 1.80898 +trainer/policy/normal/log_std Min -0.801061 +eval/num steps total 70000 +eval/num paths total 70 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0340807 +eval/Actions Std 0.900372 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50313 +time/logging (s) 0.00374622 +time/sampling batch (s) 0.282617 +time/saving (s) 0.00362152 +time/training (s) 7.42346 +time/epoch (s) 10.2166 +time/total (s) 750.192 +Epoch -931 +---------------------------------- --------------- +2022-05-10 13:23:14.364906 PDT | [0] Epoch -930 finished +---------------------------------- --------------- +epoch -930 +replay_buffer/size 999033 +trainer/num train calls 71000 +trainer/Policy Loss -19.274 +trainer/Log Pis Mean 24.5205 +trainer/Log Pis Std 13.5848 +trainer/Log Pis Max 69.8329 +trainer/Log Pis Min -7.28152 +trainer/policy/mean Mean -0.0520488 +trainer/policy/mean Std 0.899406 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.90665 +trainer/policy/normal/std Std 0.551432 +trainer/policy/normal/std Max 5.21214 +trainer/policy/normal/std Min 0.564583 +trainer/policy/normal/log_std Mean 1.04285 +trainer/policy/normal/log_std Std 0.241427 +trainer/policy/normal/log_std Max 1.65099 +trainer/policy/normal/log_std Min -0.571669 +eval/num steps total 71000 +eval/num paths total 71 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0864585 +eval/Actions Std 0.860009 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52852 +time/logging (s) 0.00368941 +time/sampling batch (s) 0.281977 +time/saving (s) 0.00348151 +time/training (s) 6.81583 +time/epoch (s) 9.6335 +time/total (s) 759.828 +Epoch -930 +---------------------------------- --------------- +2022-05-10 13:23:24.602280 PDT | [0] Epoch -929 finished +---------------------------------- --------------- +epoch -929 +replay_buffer/size 999033 +trainer/num train calls 72000 +trainer/Policy Loss -19.7385 +trainer/Log Pis Mean 24.4283 +trainer/Log Pis Std 13.4782 +trainer/Log Pis Max 68.1109 +trainer/Log Pis Min -5.44921 +trainer/policy/mean Mean -0.023832 +trainer/policy/mean Std 0.905309 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.90351 +trainer/policy/normal/std Std 0.53945 +trainer/policy/normal/std Max 5.14304 +trainer/policy/normal/std Min 0.519495 +trainer/policy/normal/log_std Mean 1.04323 +trainer/policy/normal/log_std Std 0.232579 +trainer/policy/normal/log_std Max 1.63764 +trainer/policy/normal/log_std Min -0.654898 +eval/num steps total 72000 +eval/num paths total 72 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0315628 +eval/Actions Std 0.951151 +eval/Actions Max 0.999998 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58668 +time/logging (s) 0.00383409 +time/sampling batch (s) 0.283891 +time/saving (s) 0.00351837 +time/training (s) 7.34067 +time/epoch (s) 10.2186 +time/total (s) 770.05 +Epoch -929 +---------------------------------- --------------- +2022-05-10 13:23:34.839570 PDT | [0] Epoch -928 finished +---------------------------------- --------------- +epoch -928 +replay_buffer/size 999033 +trainer/num train calls 73000 +trainer/Policy Loss -18.1411 +trainer/Log Pis Mean 23.9109 +trainer/Log Pis Std 13.0083 +trainer/Log Pis Max 67.5209 +trainer/Log Pis Min -16.1038 +trainer/policy/mean Mean -0.0192098 +trainer/policy/mean Std 0.901424 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.9045 +trainer/policy/normal/std Std 0.546679 +trainer/policy/normal/std Max 4.84044 +trainer/policy/normal/std Min 0.576018 +trainer/policy/normal/log_std Mean 1.04254 +trainer/policy/normal/log_std Std 0.238608 +trainer/policy/normal/log_std Max 1.57701 +trainer/policy/normal/log_std Min -0.551616 +eval/num steps total 73000 +eval/num paths total 73 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0710103 +eval/Actions Std 0.903717 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62008 +time/logging (s) 0.00417597 +time/sampling batch (s) 0.282112 +time/saving (s) 0.00416393 +time/training (s) 7.30815 +time/epoch (s) 10.2187 +time/total (s) 780.271 +Epoch -928 +---------------------------------- --------------- +2022-05-10 13:23:45.995942 PDT | [0] Epoch -927 finished +---------------------------------- --------------- +epoch -927 +replay_buffer/size 999033 +trainer/num train calls 74000 +trainer/Policy Loss -18.868 +trainer/Log Pis Mean 23.8769 +trainer/Log Pis Std 12.7541 +trainer/Log Pis Max 71.9846 +trainer/Log Pis Min -9.10827 +trainer/policy/mean Mean -0.0481622 +trainer/policy/mean Std 0.903772 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91308 +trainer/policy/normal/std Std 0.565417 +trainer/policy/normal/std Max 5.69492 +trainer/policy/normal/std Min 0.487869 +trainer/policy/normal/log_std Mean 1.04332 +trainer/policy/normal/log_std Std 0.251792 +trainer/policy/normal/log_std Max 1.73957 +trainer/policy/normal/log_std Min -0.717708 +eval/num steps total 74000 +eval/num paths total 74 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0198977 +eval/Actions Std 0.940603 +eval/Actions Max 0.999992 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70394 +time/logging (s) 0.00373575 +time/sampling batch (s) 0.280952 +time/saving (s) 0.00363241 +time/training (s) 8.14445 +time/epoch (s) 11.1367 +time/total (s) 791.411 +Epoch -927 +---------------------------------- --------------- +2022-05-10 13:23:57.054202 PDT | [0] Epoch -926 finished +---------------------------------- --------------- +epoch -926 +replay_buffer/size 999033 +trainer/num train calls 75000 +trainer/Policy Loss -19.754 +trainer/Log Pis Mean 24.0108 +trainer/Log Pis Std 13.8036 +trainer/Log Pis Max 77.4856 +trainer/Log Pis Min -9.91104 +trainer/policy/mean Mean -0.0511426 +trainer/policy/mean Std 0.905124 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.91829 +trainer/policy/normal/std Std 0.551579 +trainer/policy/normal/std Max 6.09764 +trainer/policy/normal/std Min 0.4915 +trainer/policy/normal/log_std Mean 1.04701 +trainer/policy/normal/log_std Std 0.240581 +trainer/policy/normal/log_std Max 1.8079 +trainer/policy/normal/log_std Min -0.710294 +eval/num steps total 75000 +eval/num paths total 75 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.375601 +eval/Actions Std 0.858653 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60547 +time/logging (s) 0.00369052 +time/sampling batch (s) 0.280273 +time/saving (s) 0.00357921 +time/training (s) 8.14634 +time/epoch (s) 11.0394 +time/total (s) 802.453 +Epoch -926 +---------------------------------- --------------- +2022-05-10 13:24:08.517920 PDT | [0] Epoch -925 finished +---------------------------------- --------------- +epoch -925 +replay_buffer/size 999033 +trainer/num train calls 76000 +trainer/Policy Loss -20.1281 +trainer/Log Pis Mean 24.254 +trainer/Log Pis Std 12.8545 +trainer/Log Pis Max 76.6595 +trainer/Log Pis Min -7.94265 +trainer/policy/mean Mean -0.0260818 +trainer/policy/mean Std 0.904895 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.91274 +trainer/policy/normal/std Std 0.53263 +trainer/policy/normal/std Max 4.88808 +trainer/policy/normal/std Min 0.539388 +trainer/policy/normal/log_std Mean 1.0469 +trainer/policy/normal/log_std Std 0.230225 +trainer/policy/normal/log_std Max 1.5868 +trainer/policy/normal/log_std Min -0.617321 +eval/num steps total 76000 +eval/num paths total 76 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.183221 +eval/Actions Std 0.911178 +eval/Actions Max 0.999916 +eval/Actions Min -0.999977 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5963 +time/logging (s) 0.00373392 +time/sampling batch (s) 0.531299 +time/saving (s) 0.00346525 +time/training (s) 8.3099 +time/epoch (s) 11.4447 +time/total (s) 813.901 +Epoch -925 +---------------------------------- --------------- +2022-05-10 13:24:18.080988 PDT | [0] Epoch -924 finished +---------------------------------- --------------- +epoch -924 +replay_buffer/size 999033 +trainer/num train calls 77000 +trainer/Policy Loss -19.9961 +trainer/Log Pis Mean 25.7321 +trainer/Log Pis Std 12.9944 +trainer/Log Pis Max 71.7057 +trainer/Log Pis Min -0.799238 +trainer/policy/mean Mean -0.0548715 +trainer/policy/mean Std 0.909219 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89735 +trainer/policy/normal/std Std 0.529964 +trainer/policy/normal/std Max 5.07057 +trainer/policy/normal/std Min 0.497808 +trainer/policy/normal/log_std Mean 1.04139 +trainer/policy/normal/log_std Std 0.232568 +trainer/policy/normal/log_std Max 1.62345 +trainer/policy/normal/log_std Min -0.697541 +eval/num steps total 77000 +eval/num paths total 77 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0675579 +eval/Actions Std 0.914713 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74849 +time/logging (s) 0.00378023 +time/sampling batch (s) 0.282988 +time/saving (s) 0.00347738 +time/training (s) 6.50536 +time/epoch (s) 9.54409 +time/total (s) 823.448 +Epoch -924 +---------------------------------- --------------- +2022-05-10 13:24:28.611682 PDT | [0] Epoch -923 finished +---------------------------------- --------------- +epoch -923 +replay_buffer/size 999033 +trainer/num train calls 78000 +trainer/Policy Loss -19.8367 +trainer/Log Pis Mean 24.6266 +trainer/Log Pis Std 13.0906 +trainer/Log Pis Max 66.5729 +trainer/Log Pis Min -2.44654 +trainer/policy/mean Mean -0.0394223 +trainer/policy/mean Std 0.903957 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87188 +trainer/policy/normal/std Std 0.528816 +trainer/policy/normal/std Max 4.78182 +trainer/policy/normal/std Min 0.61683 +trainer/policy/normal/log_std Mean 1.03163 +trainer/policy/normal/log_std Std 0.238643 +trainer/policy/normal/log_std Max 1.56482 +trainer/policy/normal/log_std Min -0.483162 +eval/num steps total 78000 +eval/num paths total 78 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.255547 +eval/Actions Std 0.843019 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6861 +time/logging (s) 0.00405021 +time/sampling batch (s) 0.533567 +time/saving (s) 0.00415141 +time/training (s) 7.28404 +time/epoch (s) 10.5119 +time/total (s) 833.962 +Epoch -923 +---------------------------------- --------------- +2022-05-10 13:24:39.282905 PDT | [0] Epoch -922 finished +---------------------------------- --------------- +epoch -922 +replay_buffer/size 999033 +trainer/num train calls 79000 +trainer/Policy Loss -18.5865 +trainer/Log Pis Mean 23.5953 +trainer/Log Pis Std 13.6033 +trainer/Log Pis Max 78.8226 +trainer/Log Pis Min -8.79383 +trainer/policy/mean Mean -0.032626 +trainer/policy/mean Std 0.897325 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.88587 +trainer/policy/normal/std Std 0.563289 +trainer/policy/normal/std Max 5.63115 +trainer/policy/normal/std Min 0.460762 +trainer/policy/normal/log_std Mean 1.03376 +trainer/policy/normal/log_std Std 0.252146 +trainer/policy/normal/log_std Max 1.72831 +trainer/policy/normal/log_std Min -0.774875 +eval/num steps total 79000 +eval/num paths total 79 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0455364 +eval/Actions Std 0.89753 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46725 +time/logging (s) 0.00385501 +time/sampling batch (s) 0.535786 +time/saving (s) 0.0035544 +time/training (s) 7.64104 +time/epoch (s) 10.6515 +time/total (s) 844.617 +Epoch -922 +---------------------------------- --------------- +2022-05-10 13:24:49.047921 PDT | [0] Epoch -921 finished +---------------------------------- --------------- +epoch -921 +replay_buffer/size 999033 +trainer/num train calls 80000 +trainer/Policy Loss -20.1103 +trainer/Log Pis Mean 26.3407 +trainer/Log Pis Std 14.4402 +trainer/Log Pis Max 84.3032 +trainer/Log Pis Min -5.4375 +trainer/policy/mean Mean -0.0483799 +trainer/policy/mean Std 0.908289 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.89943 +trainer/policy/normal/std Std 0.543337 +trainer/policy/normal/std Max 4.92508 +trainer/policy/normal/std Min 0.471982 +trainer/policy/normal/log_std Mean 1.04061 +trainer/policy/normal/log_std Std 0.241341 +trainer/policy/normal/log_std Max 1.59434 +trainer/policy/normal/log_std Min -0.750814 +eval/num steps total 80000 +eval/num paths total 80 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00983369 +eval/Actions Std 0.893625 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56649 +time/logging (s) 0.00372889 +time/sampling batch (s) 0.278176 +time/saving (s) 0.00342758 +time/training (s) 6.89442 +time/epoch (s) 9.74625 +time/total (s) 854.366 +Epoch -921 +---------------------------------- --------------- +2022-05-10 13:24:58.852707 PDT | [0] Epoch -920 finished +---------------------------------- --------------- +epoch -920 +replay_buffer/size 999033 +trainer/num train calls 81000 +trainer/Policy Loss -19.3863 +trainer/Log Pis Mean 26.0862 +trainer/Log Pis Std 13.0381 +trainer/Log Pis Max 79.5291 +trainer/Log Pis Min -5.29207 +trainer/policy/mean Mean -0.0365642 +trainer/policy/mean Std 0.905688 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.9066 +trainer/policy/normal/std Std 0.582125 +trainer/policy/normal/std Max 5.80109 +trainer/policy/normal/std Min 0.55677 +trainer/policy/normal/log_std Mean 1.03911 +trainer/policy/normal/log_std Std 0.261521 +trainer/policy/normal/log_std Max 1.75805 +trainer/policy/normal/log_std Min -0.585604 +eval/num steps total 81000 +eval/num paths total 81 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104482 +eval/Actions Std 0.906637 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66721 +time/logging (s) 0.00415848 +time/sampling batch (s) 0.278248 +time/saving (s) 0.00412538 +time/training (s) 6.8328 +time/epoch (s) 9.78654 +time/total (s) 864.156 +Epoch -920 +---------------------------------- --------------- +2022-05-10 13:25:09.124199 PDT | [0] Epoch -919 finished +---------------------------------- --------------- +epoch -919 +replay_buffer/size 999033 +trainer/num train calls 82000 +trainer/Policy Loss -19.2082 +trainer/Log Pis Mean 24.3119 +trainer/Log Pis Std 12.7434 +trainer/Log Pis Max 61.8289 +trainer/Log Pis Min -5.2845 +trainer/policy/mean Mean -0.0157748 +trainer/policy/mean Std 0.906297 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.89853 +trainer/policy/normal/std Std 0.577727 +trainer/policy/normal/std Max 4.87939 +trainer/policy/normal/std Min 0.482432 +trainer/policy/normal/log_std Mean 1.03631 +trainer/policy/normal/log_std Std 0.262736 +trainer/policy/normal/log_std Max 1.58502 +trainer/policy/normal/log_std Min -0.728916 +eval/num steps total 82000 +eval/num paths total 82 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0677525 +eval/Actions Std 0.908984 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6109 +time/logging (s) 0.00371844 +time/sampling batch (s) 0.277712 +time/saving (s) 0.00366464 +time/training (s) 7.35599 +time/epoch (s) 10.252 +time/total (s) 874.411 +Epoch -919 +---------------------------------- --------------- +2022-05-10 13:25:19.304549 PDT | [0] Epoch -918 finished +---------------------------------- --------------- +epoch -918 +replay_buffer/size 999033 +trainer/num train calls 83000 +trainer/Policy Loss -20.8092 +trainer/Log Pis Mean 23.8958 +trainer/Log Pis Std 13.092 +trainer/Log Pis Max 68.0461 +trainer/Log Pis Min -10.2346 +trainer/policy/mean Mean -0.0272483 +trainer/policy/mean Std 0.906627 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.91494 +trainer/policy/normal/std Std 0.557145 +trainer/policy/normal/std Max 5.57816 +trainer/policy/normal/std Min 0.523225 +trainer/policy/normal/log_std Mean 1.04617 +trainer/policy/normal/log_std Std 0.23639 +trainer/policy/normal/log_std Max 1.71886 +trainer/policy/normal/log_std Min -0.647744 +eval/num steps total 83000 +eval/num paths total 83 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222608 +eval/Actions Std 0.915715 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5757 +time/logging (s) 0.00379372 +time/sampling batch (s) 0.531373 +time/saving (s) 0.00346556 +time/training (s) 7.0472 +time/epoch (s) 10.1615 +time/total (s) 884.575 +Epoch -918 +---------------------------------- --------------- +2022-05-10 13:25:29.828918 PDT | [0] Epoch -917 finished +---------------------------------- --------------- +epoch -917 +replay_buffer/size 999033 +trainer/num train calls 84000 +trainer/Policy Loss -20.1128 +trainer/Log Pis Mean 25.1529 +trainer/Log Pis Std 13.9963 +trainer/Log Pis Max 96.8659 +trainer/Log Pis Min -8.94904 +trainer/policy/mean Mean -0.0391437 +trainer/policy/mean Std 0.90614 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.92715 +trainer/policy/normal/std Std 0.53785 +trainer/policy/normal/std Max 5.31966 +trainer/policy/normal/std Min 0.530974 +trainer/policy/normal/log_std Mean 1.05195 +trainer/policy/normal/log_std Std 0.229397 +trainer/policy/normal/log_std Max 1.67141 +trainer/policy/normal/log_std Min -0.633043 +eval/num steps total 84000 +eval/num paths total 84 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.131946 +eval/Actions Std 0.898974 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66842 +time/logging (s) 0.00370358 +time/sampling batch (s) 0.289729 +time/saving (s) 0.0034896 +time/training (s) 7.53957 +time/epoch (s) 10.5049 +time/total (s) 895.083 +Epoch -917 +---------------------------------- --------------- +2022-05-10 13:25:40.834732 PDT | [0] Epoch -916 finished +---------------------------------- --------------- +epoch -916 +replay_buffer/size 999033 +trainer/num train calls 85000 +trainer/Policy Loss -19.14 +trainer/Log Pis Mean 24.1971 +trainer/Log Pis Std 13.0478 +trainer/Log Pis Max 66.5733 +trainer/Log Pis Min -5.62472 +trainer/policy/mean Mean -0.0423247 +trainer/policy/mean Std 0.901188 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86448 +trainer/policy/normal/std Std 0.539268 +trainer/policy/normal/std Max 5.08738 +trainer/policy/normal/std Min 0.435541 +trainer/policy/normal/log_std Mean 1.0284 +trainer/policy/normal/log_std Std 0.241543 +trainer/policy/normal/log_std Max 1.62676 +trainer/policy/normal/log_std Min -0.831167 +eval/num steps total 85000 +eval/num paths total 85 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.199073 +eval/Actions Std 0.925856 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64818 +time/logging (s) 0.00444909 +time/sampling batch (s) 0.284503 +time/saving (s) 0.00448119 +time/training (s) 8.04561 +time/epoch (s) 10.9872 +time/total (s) 906.073 +Epoch -916 +---------------------------------- --------------- +2022-05-10 13:25:52.006331 PDT | [0] Epoch -915 finished +---------------------------------- --------------- +epoch -915 +replay_buffer/size 999033 +trainer/num train calls 86000 +trainer/Policy Loss -18.9002 +trainer/Log Pis Mean 23.547 +trainer/Log Pis Std 13.0873 +trainer/Log Pis Max 87.9046 +trainer/Log Pis Min -9.19369 +trainer/policy/mean Mean -0.0216493 +trainer/policy/mean Std 0.900683 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.90817 +trainer/policy/normal/std Std 0.584584 +trainer/policy/normal/std Max 5.18403 +trainer/policy/normal/std Min 0.465541 +trainer/policy/normal/log_std Mean 1.03926 +trainer/policy/normal/log_std Std 0.264425 +trainer/policy/normal/log_std Max 1.64558 +trainer/policy/normal/log_std Min -0.764555 +eval/num steps total 86000 +eval/num paths total 86 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.203297 +eval/Actions Std 0.8929 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65047 +time/logging (s) 0.0038043 +time/sampling batch (s) 0.287046 +time/saving (s) 0.00347934 +time/training (s) 8.20638 +time/epoch (s) 11.1512 +time/total (s) 917.228 +Epoch -915 +---------------------------------- --------------- +2022-05-10 13:26:02.375547 PDT | [0] Epoch -914 finished +---------------------------------- --------------- +epoch -914 +replay_buffer/size 999033 +trainer/num train calls 87000 +trainer/Policy Loss -18.4303 +trainer/Log Pis Mean 24.4301 +trainer/Log Pis Std 13.2077 +trainer/Log Pis Max 62.9185 +trainer/Log Pis Min -8.94891 +trainer/policy/mean Mean 0.00379956 +trainer/policy/mean Std 0.904012 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.90499 +trainer/policy/normal/std Std 0.555273 +trainer/policy/normal/std Max 5.04714 +trainer/policy/normal/std Min 0.425054 +trainer/policy/normal/log_std Mean 1.04179 +trainer/policy/normal/log_std Std 0.244606 +trainer/policy/normal/log_std Max 1.61882 +trainer/policy/normal/log_std Min -0.855538 +eval/num steps total 87000 +eval/num paths total 87 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0951146 +eval/Actions Std 0.914105 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77756 +time/logging (s) 0.00369511 +time/sampling batch (s) 0.284823 +time/saving (s) 0.00347795 +time/training (s) 7.28039 +time/epoch (s) 10.3499 +time/total (s) 927.58 +Epoch -914 +---------------------------------- --------------- +2022-05-10 13:26:12.207898 PDT | [0] Epoch -913 finished +---------------------------------- --------------- +epoch -913 +replay_buffer/size 999033 +trainer/num train calls 88000 +trainer/Policy Loss -18.196 +trainer/Log Pis Mean 23.3802 +trainer/Log Pis Std 12.5884 +trainer/Log Pis Max 66.8241 +trainer/Log Pis Min -7.81719 +trainer/policy/mean Mean -0.0462161 +trainer/policy/mean Std 0.903764 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89062 +trainer/policy/normal/std Std 0.533366 +trainer/policy/normal/std Max 4.75029 +trainer/policy/normal/std Min 0.519237 +trainer/policy/normal/log_std Mean 1.03922 +trainer/policy/normal/log_std Std 0.229921 +trainer/policy/normal/log_std Max 1.55821 +trainer/policy/normal/log_std Min -0.655395 +eval/num steps total 88000 +eval/num paths total 88 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0728231 +eval/Actions Std 0.883011 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53974 +time/logging (s) 0.00412517 +time/sampling batch (s) 0.782034 +time/saving (s) 0.00429859 +time/training (s) 6.48365 +time/epoch (s) 9.81385 +time/total (s) 937.397 +Epoch -913 +---------------------------------- --------------- +2022-05-10 13:26:22.411117 PDT | [0] Epoch -912 finished +---------------------------------- --------------- +epoch -912 +replay_buffer/size 999033 +trainer/num train calls 89000 +trainer/Policy Loss -19.4845 +trainer/Log Pis Mean 24.7861 +trainer/Log Pis Std 13.3787 +trainer/Log Pis Max 93.317 +trainer/Log Pis Min -4.85542 +trainer/policy/mean Mean -0.0671544 +trainer/policy/mean Std 0.902837 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.88295 +trainer/policy/normal/std Std 0.561947 +trainer/policy/normal/std Max 5.67374 +trainer/policy/normal/std Min 0.553707 +trainer/policy/normal/log_std Mean 1.03351 +trainer/policy/normal/log_std Std 0.24667 +trainer/policy/normal/log_std Max 1.73585 +trainer/policy/normal/log_std Min -0.59112 +eval/num steps total 89000 +eval/num paths total 89 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0322821 +eval/Actions Std 0.944821 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71415 +time/logging (s) 0.00378126 +time/sampling batch (s) 0.281951 +time/saving (s) 0.00353644 +time/training (s) 7.18009 +time/epoch (s) 10.1835 +time/total (s) 947.584 +Epoch -912 +---------------------------------- --------------- +2022-05-10 13:26:33.176308 PDT | [0] Epoch -911 finished +---------------------------------- --------------- +epoch -911 +replay_buffer/size 999033 +trainer/num train calls 90000 +trainer/Policy Loss -19.2286 +trainer/Log Pis Mean 24.2185 +trainer/Log Pis Std 13.5433 +trainer/Log Pis Max 69.5868 +trainer/Log Pis Min -10.3072 +trainer/policy/mean Mean -0.0447871 +trainer/policy/mean Std 0.905137 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86928 +trainer/policy/normal/std Std 0.551931 +trainer/policy/normal/std Max 4.94512 +trainer/policy/normal/std Min 0.49488 +trainer/policy/normal/log_std Mean 1.02924 +trainer/policy/normal/log_std Std 0.244842 +trainer/policy/normal/log_std Max 1.5984 +trainer/policy/normal/log_std Min -0.70344 +eval/num steps total 90000 +eval/num paths total 90 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101807 +eval/Actions Std 0.909636 +eval/Actions Max 0.999992 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5982 +time/logging (s) 0.00367103 +time/sampling batch (s) 0.532248 +time/saving (s) 0.00345076 +time/training (s) 7.60852 +time/epoch (s) 10.7461 +time/total (s) 958.333 +Epoch -911 +---------------------------------- --------------- +2022-05-10 13:26:42.766161 PDT | [0] Epoch -910 finished +---------------------------------- --------------- +epoch -910 +replay_buffer/size 999033 +trainer/num train calls 91000 +trainer/Policy Loss -18.9228 +trainer/Log Pis Mean 25.3978 +trainer/Log Pis Std 13.5296 +trainer/Log Pis Max 72.0574 +trainer/Log Pis Min -2.84339 +trainer/policy/mean Mean -0.0259472 +trainer/policy/mean Std 0.902071 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86331 +trainer/policy/normal/std Std 0.547785 +trainer/policy/normal/std Max 4.92919 +trainer/policy/normal/std Min 0.485563 +trainer/policy/normal/log_std Mean 1.02705 +trainer/policy/normal/log_std Std 0.246444 +trainer/policy/normal/log_std Max 1.59517 +trainer/policy/normal/log_std Min -0.722445 +eval/num steps total 91000 +eval/num paths total 91 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0675668 +eval/Actions Std 0.912916 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53112 +time/logging (s) 0.00369381 +time/sampling batch (s) 0.280681 +time/saving (s) 0.0034796 +time/training (s) 6.75203 +time/epoch (s) 9.57101 +time/total (s) 967.906 +Epoch -910 +---------------------------------- --------------- +2022-05-10 13:26:53.211330 PDT | [0] Epoch -909 finished +---------------------------------- --------------- +epoch -909 +replay_buffer/size 999033 +trainer/num train calls 92000 +trainer/Policy Loss -19.1161 +trainer/Log Pis Mean 24.3901 +trainer/Log Pis Std 13.2057 +trainer/Log Pis Max 70.0683 +trainer/Log Pis Min -9.50394 +trainer/policy/mean Mean -0.0476915 +trainer/policy/mean Std 0.905492 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.9153 +trainer/policy/normal/std Std 0.554239 +trainer/policy/normal/std Max 5.94924 +trainer/policy/normal/std Min 0.492266 +trainer/policy/normal/log_std Mean 1.04581 +trainer/policy/normal/log_std Std 0.241452 +trainer/policy/normal/log_std Max 1.78326 +trainer/policy/normal/log_std Min -0.708735 +eval/num steps total 92000 +eval/num paths total 92 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.340865 +eval/Actions Std 0.871462 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.92276 +time/logging (s) 0.00412071 +time/sampling batch (s) 0.283904 +time/saving (s) 0.00413736 +time/training (s) 7.2116 +time/epoch (s) 10.4265 +time/total (s) 978.336 +Epoch -909 +---------------------------------- --------------- +2022-05-10 13:27:03.973991 PDT | [0] Epoch -908 finished +---------------------------------- --------------- +epoch -908 +replay_buffer/size 999033 +trainer/num train calls 93000 +trainer/Policy Loss -18.2219 +trainer/Log Pis Mean 24.8898 +trainer/Log Pis Std 13.024 +trainer/Log Pis Max 72.8165 +trainer/Log Pis Min -7.03873 +trainer/policy/mean Mean -0.0337905 +trainer/policy/mean Std 0.899295 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.92723 +trainer/policy/normal/std Std 0.569078 +trainer/policy/normal/std Max 5.72068 +trainer/policy/normal/std Min 0.53552 +trainer/policy/normal/log_std Mean 1.04929 +trainer/policy/normal/log_std Std 0.242999 +trainer/policy/normal/log_std Max 1.74409 +trainer/policy/normal/log_std Min -0.624517 +eval/num steps total 93000 +eval/num paths total 93 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0752336 +eval/Actions Std 0.901668 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77698 +time/logging (s) 0.00368465 +time/sampling batch (s) 0.280645 +time/saving (s) 0.00353511 +time/training (s) 7.67816 +time/epoch (s) 10.743 +time/total (s) 989.082 +Epoch -908 +---------------------------------- --------------- +2022-05-10 13:27:13.264142 PDT | [0] Epoch -907 finished +---------------------------------- --------------- +epoch -907 +replay_buffer/size 999033 +trainer/num train calls 94000 +trainer/Policy Loss -19.1583 +trainer/Log Pis Mean 24.443 +trainer/Log Pis Std 13.2128 +trainer/Log Pis Max 59.8172 +trainer/Log Pis Min -7.45986 +trainer/policy/mean Mean -0.0248002 +trainer/policy/mean Std 0.8993 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.90342 +trainer/policy/normal/std Std 0.560346 +trainer/policy/normal/std Max 5.96199 +trainer/policy/normal/std Min 0.574648 +trainer/policy/normal/log_std Mean 1.04154 +trainer/policy/normal/log_std Std 0.240749 +trainer/policy/normal/log_std Max 1.78541 +trainer/policy/normal/log_std Min -0.553998 +eval/num steps total 94000 +eval/num paths total 94 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00280277 +eval/Actions Std 0.888466 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46703 +time/logging (s) 0.00383055 +time/sampling batch (s) 0.282702 +time/saving (s) 0.00350151 +time/training (s) 6.51426 +time/epoch (s) 9.27132 +time/total (s) 998.356 +Epoch -907 +---------------------------------- --------------- +2022-05-10 13:27:23.106041 PDT | [0] Epoch -906 finished +---------------------------------- --------------- +epoch -906 +replay_buffer/size 999033 +trainer/num train calls 95000 +trainer/Policy Loss -18.8076 +trainer/Log Pis Mean 24.7629 +trainer/Log Pis Std 13.0141 +trainer/Log Pis Max 59.6873 +trainer/Log Pis Min -6.78489 +trainer/policy/mean Mean -0.0405813 +trainer/policy/mean Std 0.903758 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83221 +trainer/policy/normal/std Std 0.564835 +trainer/policy/normal/std Max 5.06157 +trainer/policy/normal/std Min 0.510735 +trainer/policy/normal/log_std Mean 1.01336 +trainer/policy/normal/log_std Std 0.260933 +trainer/policy/normal/log_std Max 1.62168 +trainer/policy/normal/log_std Min -0.671905 +eval/num steps total 95000 +eval/num paths total 95 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111157 +eval/Actions Std 0.905753 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46121 +time/logging (s) 0.00418564 +time/sampling batch (s) 0.291122 +time/saving (s) 0.00413468 +time/training (s) 7.06192 +time/epoch (s) 9.82257 +time/total (s) 1008.18 +Epoch -906 +---------------------------------- --------------- +2022-05-10 13:27:33.910959 PDT | [0] Epoch -905 finished +---------------------------------- --------------- +epoch -905 +replay_buffer/size 999033 +trainer/num train calls 96000 +trainer/Policy Loss -21.4508 +trainer/Log Pis Mean 24.9587 +trainer/Log Pis Std 12.8781 +trainer/Log Pis Max 72.4713 +trainer/Log Pis Min -10.3046 +trainer/policy/mean Mean -0.0449052 +trainer/policy/mean Std 0.904458 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83369 +trainer/policy/normal/std Std 0.549979 +trainer/policy/normal/std Max 5.27694 +trainer/policy/normal/std Min 0.542114 +trainer/policy/normal/log_std Mean 1.01713 +trainer/policy/normal/log_std Std 0.240553 +trainer/policy/normal/log_std Max 1.66335 +trainer/policy/normal/log_std Min -0.612278 +eval/num steps total 96000 +eval/num paths total 96 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.196056 +eval/Actions Std 0.924412 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64921 +time/logging (s) 0.00367692 +time/sampling batch (s) 0.283049 +time/saving (s) 0.0035507 +time/training (s) 7.84558 +time/epoch (s) 10.7851 +time/total (s) 1018.97 +Epoch -905 +---------------------------------- --------------- +2022-05-10 13:27:44.465148 PDT | [0] Epoch -904 finished +---------------------------------- --------------- +epoch -904 +replay_buffer/size 999033 +trainer/num train calls 97000 +trainer/Policy Loss -19.6583 +trainer/Log Pis Mean 25.3816 +trainer/Log Pis Std 13.578 +trainer/Log Pis Max 69.4967 +trainer/Log Pis Min -7.03023 +trainer/policy/mean Mean -0.0598132 +trainer/policy/mean Std 0.911251 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.93781 +trainer/policy/normal/std Std 0.563723 +trainer/policy/normal/std Max 5.43068 +trainer/policy/normal/std Min 0.504835 +trainer/policy/normal/log_std Mean 1.05327 +trainer/policy/normal/log_std Std 0.241969 +trainer/policy/normal/log_std Max 1.69206 +trainer/policy/normal/log_std Min -0.683524 +eval/num steps total 97000 +eval/num paths total 97 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0419103 +eval/Actions Std 0.902558 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62819 +time/logging (s) 0.00372438 +time/sampling batch (s) 0.282269 +time/saving (s) 0.00346946 +time/training (s) 7.61742 +time/epoch (s) 10.5351 +time/total (s) 1029.51 +Epoch -904 +---------------------------------- --------------- +2022-05-10 13:27:55.841011 PDT | [0] Epoch -903 finished +---------------------------------- --------------- +epoch -903 +replay_buffer/size 999033 +trainer/num train calls 98000 +trainer/Policy Loss -18.9676 +trainer/Log Pis Mean 25.8367 +trainer/Log Pis Std 13.8956 +trainer/Log Pis Max 75.2511 +trainer/Log Pis Min -8.38732 +trainer/policy/mean Mean -0.0472755 +trainer/policy/mean Std 0.907557 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.89749 +trainer/policy/normal/std Std 0.572763 +trainer/policy/normal/std Max 5.24111 +trainer/policy/normal/std Min 0.500798 +trainer/policy/normal/log_std Mean 1.03633 +trainer/policy/normal/log_std Std 0.261295 +trainer/policy/normal/log_std Max 1.65653 +trainer/policy/normal/log_std Min -0.691553 +eval/num steps total 98000 +eval/num paths total 98 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00386493 +eval/Actions Std 0.974306 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59031 +time/logging (s) 0.00446229 +time/sampling batch (s) 0.531725 +time/saving (s) 0.00421303 +time/training (s) 8.22671 +time/epoch (s) 11.3574 +time/total (s) 1040.87 +Epoch -903 +---------------------------------- --------------- +2022-05-10 13:28:05.828070 PDT | [0] Epoch -902 finished +---------------------------------- --------------- +epoch -902 +replay_buffer/size 999033 +trainer/num train calls 99000 +trainer/Policy Loss -19.8528 +trainer/Log Pis Mean 23.5834 +trainer/Log Pis Std 13.1895 +trainer/Log Pis Max 77.4396 +trainer/Log Pis Min -9.01169 +trainer/policy/mean Mean -0.0477394 +trainer/policy/mean Std 0.904913 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.8814 +trainer/policy/normal/std Std 0.548008 +trainer/policy/normal/std Max 4.82465 +trainer/policy/normal/std Min 0.470281 +trainer/policy/normal/log_std Mean 1.03381 +trainer/policy/normal/log_std Std 0.24408 +trainer/policy/normal/log_std Max 1.57374 +trainer/policy/normal/log_std Min -0.754425 +eval/num steps total 99000 +eval/num paths total 99 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167421 +eval/Actions Std 0.871753 +eval/Actions Max 0.999994 +eval/Actions Min -0.999974 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58529 +time/logging (s) 0.00378391 +time/sampling batch (s) 0.284285 +time/saving (s) 0.00371546 +time/training (s) 7.08933 +time/epoch (s) 9.9664 +time/total (s) 1050.84 +Epoch -902 +---------------------------------- --------------- +2022-05-10 13:28:15.724463 PDT | [0] Epoch -901 finished +---------------------------------- --------------- +epoch -901 +replay_buffer/size 999033 +trainer/num train calls 100000 +trainer/Policy Loss -20.292 +trainer/Log Pis Mean 24.5281 +trainer/Log Pis Std 12.6524 +trainer/Log Pis Max 74.3459 +trainer/Log Pis Min -5.86769 +trainer/policy/mean Mean -0.0221809 +trainer/policy/mean Std 0.909232 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.88297 +trainer/policy/normal/std Std 0.574059 +trainer/policy/normal/std Max 5.20192 +trainer/policy/normal/std Min 0.46118 +trainer/policy/normal/log_std Mean 1.0311 +trainer/policy/normal/log_std Std 0.261587 +trainer/policy/normal/log_std Max 1.64903 +trainer/policy/normal/log_std Min -0.773967 +eval/num steps total 100000 +eval/num paths total 100 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.375763 +eval/Actions Std 0.822951 +eval/Actions Max 0.999991 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85768 +time/logging (s) 0.00381755 +time/sampling batch (s) 0.279426 +time/saving (s) 0.00649954 +time/training (s) 6.73018 +time/epoch (s) 9.8776 +time/total (s) 1060.72 +Epoch -901 +---------------------------------- --------------- +2022-05-10 13:28:25.898185 PDT | [0] Epoch -900 finished +---------------------------------- --------------- +epoch -900 +replay_buffer/size 999033 +trainer/num train calls 101000 +trainer/Policy Loss -18.7972 +trainer/Log Pis Mean 25.6216 +trainer/Log Pis Std 13.0555 +trainer/Log Pis Max 65.6323 +trainer/Log Pis Min -7.84558 +trainer/policy/mean Mean -0.0604618 +trainer/policy/mean Std 0.903029 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.91115 +trainer/policy/normal/std Std 0.561742 +trainer/policy/normal/std Max 5.31926 +trainer/policy/normal/std Min 0.400827 +trainer/policy/normal/log_std Mean 1.04358 +trainer/policy/normal/log_std Std 0.245993 +trainer/policy/normal/log_std Max 1.67133 +trainer/policy/normal/log_std Min -0.914225 +eval/num steps total 101000 +eval/num paths total 101 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00553888 +eval/Actions Std 0.900047 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73835 +time/logging (s) 0.00382987 +time/sampling batch (s) 0.529421 +time/saving (s) 0.00344979 +time/training (s) 6.87972 +time/epoch (s) 10.1548 +time/total (s) 1070.88 +Epoch -900 +---------------------------------- --------------- +2022-05-10 13:28:35.769805 PDT | [0] Epoch -899 finished +---------------------------------- --------------- +epoch -899 +replay_buffer/size 999033 +trainer/num train calls 102000 +trainer/Policy Loss -19.8224 +trainer/Log Pis Mean 24.2642 +trainer/Log Pis Std 12.8934 +trainer/Log Pis Max 77.0606 +trainer/Log Pis Min -5.63246 +trainer/policy/mean Mean -0.0283063 +trainer/policy/mean Std 0.907084 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85971 +trainer/policy/normal/std Std 0.53317 +trainer/policy/normal/std Max 5.28132 +trainer/policy/normal/std Min 0.447472 +trainer/policy/normal/log_std Mean 1.02673 +trainer/policy/normal/log_std Std 0.242353 +trainer/policy/normal/log_std Max 1.66418 +trainer/policy/normal/log_std Min -0.804141 +eval/num steps total 102000 +eval/num paths total 102 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127852 +eval/Actions Std 0.881596 +eval/Actions Max 0.99998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65105 +time/logging (s) 0.00412278 +time/sampling batch (s) 0.283604 +time/saving (s) 0.00407927 +time/training (s) 6.90991 +time/epoch (s) 9.85276 +time/total (s) 1080.73 +Epoch -899 +---------------------------------- --------------- +2022-05-10 13:28:48.773032 PDT | [0] Epoch -898 finished +---------------------------------- --------------- +epoch -898 +replay_buffer/size 999033 +trainer/num train calls 103000 +trainer/Policy Loss -18.8577 +trainer/Log Pis Mean 24.5871 +trainer/Log Pis Std 12.7787 +trainer/Log Pis Max 73.4107 +trainer/Log Pis Min -9.5822 +trainer/policy/mean Mean -0.0464077 +trainer/policy/mean Std 0.906923 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.92929 +trainer/policy/normal/std Std 0.546961 +trainer/policy/normal/std Max 6.10565 +trainer/policy/normal/std Min 0.56379 +trainer/policy/normal/log_std Mean 1.05174 +trainer/policy/normal/log_std Std 0.234967 +trainer/policy/normal/log_std Max 1.80922 +trainer/policy/normal/log_std Min -0.573073 +eval/num steps total 103000 +eval/num paths total 103 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.150085 +eval/Actions Std 0.938877 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51161 +time/logging (s) 0.00383507 +time/sampling batch (s) 0.534241 +time/saving (s) 0.00364499 +time/training (s) 9.9302 +time/epoch (s) 12.9835 +time/total (s) 1093.72 +Epoch -898 +---------------------------------- --------------- +2022-05-10 13:28:59.346449 PDT | [0] Epoch -897 finished +---------------------------------- --------------- +epoch -897 +replay_buffer/size 999033 +trainer/num train calls 104000 +trainer/Policy Loss -20.1723 +trainer/Log Pis Mean 24.4947 +trainer/Log Pis Std 13.3096 +trainer/Log Pis Max 61.5334 +trainer/Log Pis Min -8.61767 +trainer/policy/mean Mean -0.0578275 +trainer/policy/mean Std 0.902856 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85188 +trainer/policy/normal/std Std 0.547805 +trainer/policy/normal/std Max 5.14935 +trainer/policy/normal/std Min 0.488363 +trainer/policy/normal/log_std Mean 1.0233 +trainer/policy/normal/log_std Std 0.244342 +trainer/policy/normal/log_std Max 1.63887 +trainer/policy/normal/log_std Min -0.716696 +eval/num steps total 104000 +eval/num paths total 104 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00861545 +eval/Actions Std 0.898032 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55536 +time/logging (s) 0.00369172 +time/sampling batch (s) 0.279498 +time/saving (s) 0.00345354 +time/training (s) 7.71208 +time/epoch (s) 10.5541 +time/total (s) 1104.28 +Epoch -897 +---------------------------------- --------------- +2022-05-10 13:29:09.779471 PDT | [0] Epoch -896 finished +---------------------------------- --------------- +epoch -896 +replay_buffer/size 999033 +trainer/num train calls 105000 +trainer/Policy Loss -17.9001 +trainer/Log Pis Mean 23.9757 +trainer/Log Pis Std 13.2418 +trainer/Log Pis Max 75.5773 +trainer/Log Pis Min -7.4456 +trainer/policy/mean Mean -0.0260049 +trainer/policy/mean Std 0.903008 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.93047 +trainer/policy/normal/std Std 0.564069 +trainer/policy/normal/std Max 5.61127 +trainer/policy/normal/std Min 0.381602 +trainer/policy/normal/log_std Mean 1.0497 +trainer/policy/normal/log_std Std 0.250256 +trainer/policy/normal/log_std Max 1.72478 +trainer/policy/normal/log_std Min -0.963378 +eval/num steps total 105000 +eval/num paths total 105 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135379 +eval/Actions Std 0.879445 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56316 +time/logging (s) 0.00365809 +time/sampling batch (s) 0.532266 +time/saving (s) 0.00346667 +time/training (s) 7.31129 +time/epoch (s) 10.4138 +time/total (s) 1114.69 +Epoch -896 +---------------------------------- --------------- +2022-05-10 13:29:20.500557 PDT | [0] Epoch -895 finished +---------------------------------- --------------- +epoch -895 +replay_buffer/size 999033 +trainer/num train calls 106000 +trainer/Policy Loss -20.6268 +trainer/Log Pis Mean 24.6271 +trainer/Log Pis Std 13.2765 +trainer/Log Pis Max 63.3084 +trainer/Log Pis Min -8.98444 +trainer/policy/mean Mean -0.0370805 +trainer/policy/mean Std 0.904352 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80362 +trainer/policy/normal/std Std 0.561073 +trainer/policy/normal/std Max 5.12324 +trainer/policy/normal/std Min 0.549818 +trainer/policy/normal/log_std Mean 1.00349 +trainer/policy/normal/log_std Std 0.258879 +trainer/policy/normal/log_std Max 1.63379 +trainer/policy/normal/log_std Min -0.598168 +eval/num steps total 106000 +eval/num paths total 106 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0979771 +eval/Actions Std 0.901912 +eval/Actions Max 0.999984 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68403 +time/logging (s) 0.00372769 +time/sampling batch (s) 0.28368 +time/saving (s) 0.0035007 +time/training (s) 7.72694 +time/epoch (s) 10.7019 +time/total (s) 1125.4 +Epoch -895 +---------------------------------- --------------- +2022-05-10 13:29:31.274355 PDT | [0] Epoch -894 finished +---------------------------------- --------------- +epoch -894 +replay_buffer/size 999033 +trainer/num train calls 107000 +trainer/Policy Loss -19.5842 +trainer/Log Pis Mean 24.1673 +trainer/Log Pis Std 13.5703 +trainer/Log Pis Max 68.6174 +trainer/Log Pis Min -7.97576 +trainer/policy/mean Mean -0.0624041 +trainer/policy/mean Std 0.900894 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.95537 +trainer/policy/normal/std Std 0.570671 +trainer/policy/normal/std Max 6.34356 +trainer/policy/normal/std Min 0.590816 +trainer/policy/normal/log_std Mean 1.05937 +trainer/policy/normal/log_std Std 0.239874 +trainer/policy/normal/log_std Max 1.84744 +trainer/policy/normal/log_std Min -0.52625 +eval/num steps total 107000 +eval/num paths total 107 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0732316 +eval/Actions Std 0.917661 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77085 +time/logging (s) 0.004403 +time/sampling batch (s) 0.28546 +time/saving (s) 0.00423274 +time/training (s) 7.69004 +time/epoch (s) 10.755 +time/total (s) 1136.15 +Epoch -894 +---------------------------------- --------------- +2022-05-10 13:29:41.349864 PDT | [0] Epoch -893 finished +---------------------------------- --------------- +epoch -893 +replay_buffer/size 999033 +trainer/num train calls 108000 +trainer/Policy Loss -19.8848 +trainer/Log Pis Mean 24.6563 +trainer/Log Pis Std 13.3073 +trainer/Log Pis Max 73.5081 +trainer/Log Pis Min -2.92814 +trainer/policy/mean Mean -0.0346158 +trainer/policy/mean Std 0.905459 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88644 +trainer/policy/normal/std Std 0.538368 +trainer/policy/normal/std Max 6.23494 +trainer/policy/normal/std Min 0.440142 +trainer/policy/normal/log_std Mean 1.03717 +trainer/policy/normal/log_std Std 0.234186 +trainer/policy/normal/log_std Max 1.83017 +trainer/policy/normal/log_std Min -0.820658 +eval/num steps total 108000 +eval/num paths total 108 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.255673 +eval/Actions Std 0.842011 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79056 +time/logging (s) 0.00376346 +time/sampling batch (s) 0.532519 +time/saving (s) 0.00362247 +time/training (s) 6.72465 +time/epoch (s) 10.0551 +time/total (s) 1146.21 +Epoch -893 +---------------------------------- --------------- +2022-05-10 13:29:52.311628 PDT | [0] Epoch -892 finished +---------------------------------- --------------- +epoch -892 +replay_buffer/size 999033 +trainer/num train calls 109000 +trainer/Policy Loss -19.9077 +trainer/Log Pis Mean 24.5439 +trainer/Log Pis Std 12.9169 +trainer/Log Pis Max 76.5211 +trainer/Log Pis Min -7.12206 +trainer/policy/mean Mean -0.0332165 +trainer/policy/mean Std 0.90202 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.90396 +trainer/policy/normal/std Std 0.590644 +trainer/policy/normal/std Max 5.37717 +trainer/policy/normal/std Min 0.491561 +trainer/policy/normal/log_std Mean 1.03748 +trainer/policy/normal/log_std Std 0.265647 +trainer/policy/normal/log_std Max 1.68216 +trainer/policy/normal/log_std Min -0.71017 +eval/num steps total 109000 +eval/num paths total 109 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136204 +eval/Actions Std 0.883783 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57955 +time/logging (s) 0.00367691 +time/sampling batch (s) 0.778727 +time/saving (s) 0.0033932 +time/training (s) 7.57734 +time/epoch (s) 10.9427 +time/total (s) 1157.16 +Epoch -892 +---------------------------------- --------------- +2022-05-10 13:30:02.647250 PDT | [0] Epoch -891 finished +---------------------------------- --------------- +epoch -891 +replay_buffer/size 999033 +trainer/num train calls 110000 +trainer/Policy Loss -19.5944 +trainer/Log Pis Mean 23.8071 +trainer/Log Pis Std 12.4459 +trainer/Log Pis Max 73.2721 +trainer/Log Pis Min -12.8651 +trainer/policy/mean Mean -0.0313378 +trainer/policy/mean Std 0.902623 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85082 +trainer/policy/normal/std Std 0.57803 +trainer/policy/normal/std Max 6.41318 +trainer/policy/normal/std Min 0.52922 +trainer/policy/normal/log_std Mean 1.01853 +trainer/policy/normal/log_std Std 0.269177 +trainer/policy/normal/log_std Max 1.85835 +trainer/policy/normal/log_std Min -0.636351 +eval/num steps total 110000 +eval/num paths total 110 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0273268 +eval/Actions Std 0.898898 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74476 +time/logging (s) 0.00677274 +time/sampling batch (s) 0.279864 +time/saving (s) 0.0052961 +time/training (s) 7.28321 +time/epoch (s) 10.3199 +time/total (s) 1167.48 +Epoch -891 +---------------------------------- --------------- +2022-05-10 13:30:12.918537 PDT | [0] Epoch -890 finished +---------------------------------- --------------- +epoch -890 +replay_buffer/size 999033 +trainer/num train calls 111000 +trainer/Policy Loss -19.7075 +trainer/Log Pis Mean 24.0919 +trainer/Log Pis Std 13.2906 +trainer/Log Pis Max 73.5132 +trainer/Log Pis Min -9.90425 +trainer/policy/mean Mean -0.0300347 +trainer/policy/mean Std 0.90179 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.91763 +trainer/policy/normal/std Std 0.569351 +trainer/policy/normal/std Max 5.15267 +trainer/policy/normal/std Min 0.422081 +trainer/policy/normal/log_std Mean 1.04413 +trainer/policy/normal/log_std Std 0.25674 +trainer/policy/normal/log_std Max 1.63951 +trainer/policy/normal/log_std Min -0.862557 +eval/num steps total 110558 +eval/num paths total 111 +eval/path length Mean 558 +eval/path length Std 0 +eval/path length Max 558 +eval/path length Min 558 +eval/Rewards Mean 0.00179211 +eval/Rewards Std 0.0422954 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.010221 +eval/Actions Std 0.899444 +eval/Actions Max 0.999994 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.76815 +time/logging (s) 0.00254266 +time/sampling batch (s) 0.529667 +time/saving (s) 0.00344067 +time/training (s) 6.94378 +time/epoch (s) 10.2476 +time/total (s) 1177.73 +Epoch -890 +---------------------------------- --------------- +2022-05-10 13:30:23.001455 PDT | [0] Epoch -889 finished +---------------------------------- --------------- +epoch -889 +replay_buffer/size 999033 +trainer/num train calls 112000 +trainer/Policy Loss -19.7473 +trainer/Log Pis Mean 24.4104 +trainer/Log Pis Std 13.424 +trainer/Log Pis Max 78.4587 +trainer/Log Pis Min -8.109 +trainer/policy/mean Mean -0.0628191 +trainer/policy/mean Std 0.901222 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81228 +trainer/policy/normal/std Std 0.542879 +trainer/policy/normal/std Max 4.71189 +trainer/policy/normal/std Min 0.466445 +trainer/policy/normal/log_std Mean 1.00853 +trainer/policy/normal/log_std Std 0.249106 +trainer/policy/normal/log_std Max 1.55009 +trainer/policy/normal/log_std Min -0.762615 +eval/num steps total 111558 +eval/num paths total 112 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0415015 +eval/Actions Std 0.937647 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76664 +time/logging (s) 0.00437223 +time/sampling batch (s) 0.280709 +time/saving (s) 0.00420663 +time/training (s) 7.00988 +time/epoch (s) 10.0658 +time/total (s) 1187.8 +Epoch -889 +---------------------------------- --------------- +2022-05-10 13:30:33.208855 PDT | [0] Epoch -888 finished +---------------------------------- --------------- +epoch -888 +replay_buffer/size 999033 +trainer/num train calls 113000 +trainer/Policy Loss -17.9682 +trainer/Log Pis Mean 24.4961 +trainer/Log Pis Std 13.0123 +trainer/Log Pis Max 69.2159 +trainer/Log Pis Min -7.07695 +trainer/policy/mean Mean -0.0139736 +trainer/policy/mean Std 0.899815 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.87798 +trainer/policy/normal/std Std 0.56758 +trainer/policy/normal/std Max 5.30504 +trainer/policy/normal/std Min 0.486608 +trainer/policy/normal/log_std Mean 1.03032 +trainer/policy/normal/log_std Std 0.256481 +trainer/policy/normal/log_std Max 1.66866 +trainer/policy/normal/log_std Min -0.720296 +eval/num steps total 112558 +eval/num paths total 113 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.229735 +eval/Actions Std 0.839339 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54848 +time/logging (s) 0.00376817 +time/sampling batch (s) 0.280331 +time/saving (s) 0.00360327 +time/training (s) 7.35102 +time/epoch (s) 10.1872 +time/total (s) 1197.99 +Epoch -888 +---------------------------------- --------------- +2022-05-10 13:30:43.049992 PDT | [0] Epoch -887 finished +---------------------------------- --------------- +epoch -887 +replay_buffer/size 999033 +trainer/num train calls 114000 +trainer/Policy Loss -19.9595 +trainer/Log Pis Mean 23.9785 +trainer/Log Pis Std 12.8983 +trainer/Log Pis Max 68.8588 +trainer/Log Pis Min -6.92981 +trainer/policy/mean Mean -0.027621 +trainer/policy/mean Std 0.901614 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84678 +trainer/policy/normal/std Std 0.55894 +trainer/policy/normal/std Max 5.01319 +trainer/policy/normal/std Min 0.447745 +trainer/policy/normal/log_std Mean 1.01982 +trainer/policy/normal/log_std Std 0.253852 +trainer/policy/normal/log_std Max 1.61207 +trainer/policy/normal/log_std Min -0.803532 +eval/num steps total 113558 +eval/num paths total 114 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122094 +eval/Actions Std 0.905571 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50816 +time/logging (s) 0.00391335 +time/sampling batch (s) 0.531998 +time/saving (s) 0.00367248 +time/training (s) 6.77438 +time/epoch (s) 9.82212 +time/total (s) 1207.82 +Epoch -887 +---------------------------------- --------------- +2022-05-10 13:30:53.036647 PDT | [0] Epoch -886 finished +---------------------------------- --------------- +epoch -886 +replay_buffer/size 999033 +trainer/num train calls 115000 +trainer/Policy Loss -18.9085 +trainer/Log Pis Mean 24.8727 +trainer/Log Pis Std 13.638 +trainer/Log Pis Max 69.4061 +trainer/Log Pis Min -5.82815 +trainer/policy/mean Mean -0.0383266 +trainer/policy/mean Std 0.900441 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.90607 +trainer/policy/normal/std Std 0.559013 +trainer/policy/normal/std Max 5.69066 +trainer/policy/normal/std Min 0.46008 +trainer/policy/normal/log_std Mean 1.04221 +trainer/policy/normal/log_std Std 0.243395 +trainer/policy/normal/log_std Max 1.73883 +trainer/policy/normal/log_std Min -0.776355 +eval/num steps total 114558 +eval/num paths total 115 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.080369 +eval/Actions Std 0.909248 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71947 +time/logging (s) 0.00368023 +time/sampling batch (s) 0.293815 +time/saving (s) 0.00339992 +time/training (s) 6.94615 +time/epoch (s) 9.96651 +time/total (s) 1217.79 +Epoch -886 +---------------------------------- --------------- +2022-05-10 13:31:03.647961 PDT | [0] Epoch -885 finished +---------------------------------- --------------- +epoch -885 +replay_buffer/size 999033 +trainer/num train calls 116000 +trainer/Policy Loss -19.9675 +trainer/Log Pis Mean 24.3442 +trainer/Log Pis Std 13.1594 +trainer/Log Pis Max 64.2 +trainer/Log Pis Min -4.27738 +trainer/policy/mean Mean -0.0349373 +trainer/policy/mean Std 0.906332 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.87509 +trainer/policy/normal/std Std 0.529269 +trainer/policy/normal/std Max 5.02388 +trainer/policy/normal/std Min 0.524142 +trainer/policy/normal/log_std Mean 1.03411 +trainer/policy/normal/log_std Std 0.228231 +trainer/policy/normal/log_std Max 1.6142 +trainer/policy/normal/log_std Min -0.645993 +eval/num steps total 115558 +eval/num paths total 116 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0858888 +eval/Actions Std 0.914416 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63122 +time/logging (s) 0.00368776 +time/sampling batch (s) 0.284731 +time/saving (s) 0.00345632 +time/training (s) 7.66886 +time/epoch (s) 10.592 +time/total (s) 1228.38 +Epoch -885 +---------------------------------- --------------- +2022-05-10 13:31:13.367459 PDT | [0] Epoch -884 finished +---------------------------------- --------------- +epoch -884 +replay_buffer/size 999033 +trainer/num train calls 117000 +trainer/Policy Loss -20.1361 +trainer/Log Pis Mean 24.2516 +trainer/Log Pis Std 12.7651 +trainer/Log Pis Max 66.3003 +trainer/Log Pis Min -12.5772 +trainer/policy/mean Mean -0.0303329 +trainer/policy/mean Std 0.905087 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.9056 +trainer/policy/normal/std Std 0.596549 +trainer/policy/normal/std Max 5.06282 +trainer/policy/normal/std Min 0.439586 +trainer/policy/normal/log_std Mean 1.03678 +trainer/policy/normal/log_std Std 0.272944 +trainer/policy/normal/log_std Max 1.62192 +trainer/policy/normal/log_std Min -0.821921 +eval/num steps total 116558 +eval/num paths total 117 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.272214 +eval/Actions Std 0.785142 +eval/Actions Max 0.999985 +eval/Actions Min -0.999958 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85135 +time/logging (s) 0.00370865 +time/sampling batch (s) 0.533565 +time/saving (s) 0.00349864 +time/training (s) 6.30816 +time/epoch (s) 9.70028 +time/total (s) 1238.08 +Epoch -884 +---------------------------------- --------------- +2022-05-10 13:31:24.404142 PDT | [0] Epoch -883 finished +---------------------------------- --------------- +epoch -883 +replay_buffer/size 999033 +trainer/num train calls 118000 +trainer/Policy Loss -19.0897 +trainer/Log Pis Mean 25.2058 +trainer/Log Pis Std 13.1252 +trainer/Log Pis Max 70.8615 +trainer/Log Pis Min -4.76573 +trainer/policy/mean Mean -0.0248801 +trainer/policy/mean Std 0.901855 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85808 +trainer/policy/normal/std Std 0.536187 +trainer/policy/normal/std Max 4.79295 +trainer/policy/normal/std Min 0.557673 +trainer/policy/normal/log_std Mean 1.02689 +trainer/policy/normal/log_std Std 0.236002 +trainer/policy/normal/log_std Max 1.56715 +trainer/policy/normal/log_std Min -0.583983 +eval/num steps total 117558 +eval/num paths total 118 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.125799 +eval/Actions Std 0.860089 +eval/Actions Max 0.999951 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77806 +time/logging (s) 0.00404215 +time/sampling batch (s) 0.280429 +time/saving (s) 0.00410434 +time/training (s) 7.95124 +time/epoch (s) 11.0179 +time/total (s) 1249.1 +Epoch -883 +---------------------------------- --------------- +2022-05-10 13:31:34.630216 PDT | [0] Epoch -882 finished +---------------------------------- --------------- +epoch -882 +replay_buffer/size 999033 +trainer/num train calls 119000 +trainer/Policy Loss -18.4665 +trainer/Log Pis Mean 23.4713 +trainer/Log Pis Std 12.8477 +trainer/Log Pis Max 62.9029 +trainer/Log Pis Min -8.80446 +trainer/policy/mean Mean -0.0456742 +trainer/policy/mean Std 0.901555 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87762 +trainer/policy/normal/std Std 0.581638 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.532272 +trainer/policy/normal/log_std Mean 1.02929 +trainer/policy/normal/log_std Std 0.259845 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.6306 +eval/num steps total 118558 +eval/num paths total 119 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.170432 +eval/Actions Std 0.895501 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53707 +time/logging (s) 0.00376046 +time/sampling batch (s) 0.280483 +time/saving (s) 0.00346408 +time/training (s) 7.38172 +time/epoch (s) 10.2065 +time/total (s) 1259.31 +Epoch -882 +---------------------------------- --------------- +2022-05-10 13:31:44.411751 PDT | [0] Epoch -881 finished +---------------------------------- --------------- +epoch -881 +replay_buffer/size 999033 +trainer/num train calls 120000 +trainer/Policy Loss -18.9865 +trainer/Log Pis Mean 24.6245 +trainer/Log Pis Std 13.4108 +trainer/Log Pis Max 71.9883 +trainer/Log Pis Min -8.70286 +trainer/policy/mean Mean -0.00956553 +trainer/policy/mean Std 0.902774 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86932 +trainer/policy/normal/std Std 0.584411 +trainer/policy/normal/std Max 5.75999 +trainer/policy/normal/std Min 0.371168 +trainer/policy/normal/log_std Mean 1.02568 +trainer/policy/normal/log_std Std 0.263546 +trainer/policy/normal/log_std Max 1.75094 +trainer/policy/normal/log_std Min -0.991102 +eval/num steps total 119558 +eval/num paths total 120 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.236739 +eval/Actions Std 0.880533 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43877 +time/logging (s) 0.00369049 +time/sampling batch (s) 0.281639 +time/saving (s) 0.00342385 +time/training (s) 7.03469 +time/epoch (s) 9.76222 +time/total (s) 1269.08 +Epoch -881 +---------------------------------- --------------- +2022-05-10 13:31:54.634299 PDT | [0] Epoch -880 finished +---------------------------------- --------------- +epoch -880 +replay_buffer/size 999033 +trainer/num train calls 121000 +trainer/Policy Loss -20.4359 +trainer/Log Pis Mean 24.0564 +trainer/Log Pis Std 12.7341 +trainer/Log Pis Max 68.8578 +trainer/Log Pis Min -9.99647 +trainer/policy/mean Mean -0.0482422 +trainer/policy/mean Std 0.908791 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81667 +trainer/policy/normal/std Std 0.546237 +trainer/policy/normal/std Max 5.04137 +trainer/policy/normal/std Min 0.521158 +trainer/policy/normal/log_std Mean 1.01081 +trainer/policy/normal/log_std Std 0.242949 +trainer/policy/normal/log_std Max 1.61768 +trainer/policy/normal/log_std Min -0.651702 +eval/num steps total 120558 +eval/num paths total 121 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.406097 +eval/Actions Std 0.870779 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60124 +time/logging (s) 0.00369832 +time/sampling batch (s) 0.279691 +time/saving (s) 0.00345046 +time/training (s) 7.31548 +time/epoch (s) 10.2036 +time/total (s) 1279.28 +Epoch -880 +---------------------------------- --------------- +2022-05-10 13:32:04.938217 PDT | [0] Epoch -879 finished +---------------------------------- --------------- +epoch -879 +replay_buffer/size 999033 +trainer/num train calls 122000 +trainer/Policy Loss -18.8599 +trainer/Log Pis Mean 24.719 +trainer/Log Pis Std 13.1401 +trainer/Log Pis Max 75.296 +trainer/Log Pis Min -2.31288 +trainer/policy/mean Mean -0.0358553 +trainer/policy/mean Std 0.903336 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87755 +trainer/policy/normal/std Std 0.565434 +trainer/policy/normal/std Max 5.2908 +trainer/policy/normal/std Min 0.477212 +trainer/policy/normal/log_std Mean 1.03042 +trainer/policy/normal/log_std Std 0.25481 +trainer/policy/normal/log_std Max 1.66597 +trainer/policy/normal/log_std Min -0.739794 +eval/num steps total 121558 +eval/num paths total 122 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.013917 +eval/Actions Std 0.971145 +eval/Actions Max 0.999999 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72201 +time/logging (s) 0.00364895 +time/sampling batch (s) 0.28803 +time/saving (s) 0.00346302 +time/training (s) 7.26725 +time/epoch (s) 10.2844 +time/total (s) 1289.57 +Epoch -879 +---------------------------------- --------------- +2022-05-10 13:32:15.213957 PDT | [0] Epoch -878 finished +---------------------------------- --------------- +epoch -878 +replay_buffer/size 999033 +trainer/num train calls 123000 +trainer/Policy Loss -19.1902 +trainer/Log Pis Mean 25.2527 +trainer/Log Pis Std 13.5186 +trainer/Log Pis Max 62.0958 +trainer/Log Pis Min -10.4475 +trainer/policy/mean Mean -0.0499528 +trainer/policy/mean Std 0.90928 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.93558 +trainer/policy/normal/std Std 0.593132 +trainer/policy/normal/std Max 5.72157 +trainer/policy/normal/std Min 0.479403 +trainer/policy/normal/log_std Mean 1.04874 +trainer/policy/normal/log_std Std 0.263506 +trainer/policy/normal/log_std Max 1.74424 +trainer/policy/normal/log_std Min -0.735215 +eval/num steps total 122558 +eval/num paths total 123 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0475734 +eval/Actions Std 0.926485 +eval/Actions Max 0.999998 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70165 +time/logging (s) 0.00419274 +time/sampling batch (s) 0.282179 +time/saving (s) 0.00416821 +time/training (s) 7.26496 +time/epoch (s) 10.2572 +time/total (s) 1299.83 +Epoch -878 +---------------------------------- --------------- +2022-05-10 13:32:25.559561 PDT | [0] Epoch -877 finished +---------------------------------- ---------------- +epoch -877 +replay_buffer/size 999033 +trainer/num train calls 124000 +trainer/Policy Loss -18.9421 +trainer/Log Pis Mean 24.2278 +trainer/Log Pis Std 12.969 +trainer/Log Pis Max 75.2585 +trainer/Log Pis Min -5.63214 +trainer/policy/mean Mean -0.0293882 +trainer/policy/mean Std 0.904722 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.89812 +trainer/policy/normal/std Std 0.572341 +trainer/policy/normal/std Max 5.22536 +trainer/policy/normal/std Min 0.482033 +trainer/policy/normal/log_std Mean 1.03736 +trainer/policy/normal/log_std Std 0.255575 +trainer/policy/normal/log_std Max 1.65352 +trainer/policy/normal/log_std Min -0.729743 +eval/num steps total 123558 +eval/num paths total 124 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000105366 +eval/Actions Std 0.901319 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53413 +time/logging (s) 0.00378752 +time/sampling batch (s) 0.283799 +time/saving (s) 0.00355236 +time/training (s) 7.50011 +time/epoch (s) 10.3254 +time/total (s) 1310.16 +Epoch -877 +---------------------------------- ---------------- +2022-05-10 13:32:36.229263 PDT | [0] Epoch -876 finished +---------------------------------- --------------- +epoch -876 +replay_buffer/size 999033 +trainer/num train calls 125000 +trainer/Policy Loss -19.8845 +trainer/Log Pis Mean 24.4219 +trainer/Log Pis Std 13.6638 +trainer/Log Pis Max 74.6089 +trainer/Log Pis Min -9.51374 +trainer/policy/mean Mean -0.0168101 +trainer/policy/mean Std 0.909015 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.88794 +trainer/policy/normal/std Std 0.56744 +trainer/policy/normal/std Max 4.9354 +trainer/policy/normal/std Min 0.50936 +trainer/policy/normal/log_std Mean 1.03413 +trainer/policy/normal/log_std Std 0.253633 +trainer/policy/normal/log_std Max 1.59643 +trainer/policy/normal/log_std Min -0.6746 +eval/num steps total 124558 +eval/num paths total 125 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.386616 +eval/Actions Std 0.851998 +eval/Actions Max 0.999991 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46725 +time/logging (s) 0.00382432 +time/sampling batch (s) 1.03046 +time/saving (s) 0.00351006 +time/training (s) 7.14562 +time/epoch (s) 10.6507 +time/total (s) 1320.81 +Epoch -876 +---------------------------------- --------------- +2022-05-10 13:32:46.332228 PDT | [0] Epoch -875 finished +---------------------------------- --------------- +epoch -875 +replay_buffer/size 999033 +trainer/num train calls 126000 +trainer/Policy Loss -19.1788 +trainer/Log Pis Mean 24.8553 +trainer/Log Pis Std 13.2504 +trainer/Log Pis Max 74.569 +trainer/Log Pis Min -10.2871 +trainer/policy/mean Mean -0.0620116 +trainer/policy/mean Std 0.903127 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86956 +trainer/policy/normal/std Std 0.56858 +trainer/policy/normal/std Max 5.48028 +trainer/policy/normal/std Min 0.469052 +trainer/policy/normal/log_std Mean 1.02731 +trainer/policy/normal/log_std Std 0.25622 +trainer/policy/normal/log_std Max 1.70116 +trainer/policy/normal/log_std Min -0.757041 +eval/num steps total 125558 +eval/num paths total 126 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100248 +eval/Actions Std 0.915863 +eval/Actions Max 0.999992 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47936 +time/logging (s) 0.00371587 +time/sampling batch (s) 0.532351 +time/saving (s) 0.00343666 +time/training (s) 7.0647 +time/epoch (s) 10.0836 +time/total (s) 1330.9 +Epoch -875 +---------------------------------- --------------- +2022-05-10 13:32:56.383037 PDT | [0] Epoch -874 finished +---------------------------------- --------------- +epoch -874 +replay_buffer/size 999033 +trainer/num train calls 127000 +trainer/Policy Loss -18.2351 +trainer/Log Pis Mean 25.7429 +trainer/Log Pis Std 13.6888 +trainer/Log Pis Max 68.6263 +trainer/Log Pis Min -5.69946 +trainer/policy/mean Mean -0.0403001 +trainer/policy/mean Std 0.905799 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.95803 +trainer/policy/normal/std Std 0.582812 +trainer/policy/normal/std Max 5.55489 +trainer/policy/normal/std Min 0.558259 +trainer/policy/normal/log_std Mean 1.05881 +trainer/policy/normal/log_std Std 0.248193 +trainer/policy/normal/log_std Max 1.71468 +trainer/policy/normal/log_std Min -0.582932 +eval/num steps total 126558 +eval/num paths total 127 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0879282 +eval/Actions Std 0.911217 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53946 +time/logging (s) 0.0037002 +time/sampling batch (s) 0.281107 +time/saving (s) 0.00343857 +time/training (s) 7.20392 +time/epoch (s) 10.0316 +time/total (s) 1340.93 +Epoch -874 +---------------------------------- --------------- +2022-05-10 13:33:06.214497 PDT | [0] Epoch -873 finished +---------------------------------- --------------- +epoch -873 +replay_buffer/size 999033 +trainer/num train calls 128000 +trainer/Policy Loss -19.7909 +trainer/Log Pis Mean 23.6254 +trainer/Log Pis Std 12.8264 +trainer/Log Pis Max 65.3193 +trainer/Log Pis Min -8.82375 +trainer/policy/mean Mean -0.0388005 +trainer/policy/mean Std 0.906792 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86346 +trainer/policy/normal/std Std 0.562184 +trainer/policy/normal/std Max 5.61171 +trainer/policy/normal/std Min 0.4959 +trainer/policy/normal/log_std Mean 1.02638 +trainer/policy/normal/log_std Std 0.248447 +trainer/policy/normal/log_std Max 1.72486 +trainer/policy/normal/log_std Min -0.701381 +eval/num steps total 127558 +eval/num paths total 128 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.190616 +eval/Actions Std 0.951063 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5043 +time/logging (s) 0.00388682 +time/sampling batch (s) 0.288401 +time/saving (s) 0.00361031 +time/training (s) 7.01185 +time/epoch (s) 9.81205 +time/total (s) 1350.75 +Epoch -873 +---------------------------------- --------------- +2022-05-10 13:33:16.014887 PDT | [0] Epoch -872 finished +---------------------------------- --------------- +epoch -872 +replay_buffer/size 999033 +trainer/num train calls 129000 +trainer/Policy Loss -20.4117 +trainer/Log Pis Mean 24.7539 +trainer/Log Pis Std 12.5887 +trainer/Log Pis Max 66.3011 +trainer/Log Pis Min -7.83199 +trainer/policy/mean Mean -0.0294145 +trainer/policy/mean Std 0.908599 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85667 +trainer/policy/normal/std Std 0.569969 +trainer/policy/normal/std Max 6.85972 +trainer/policy/normal/std Min 0.50741 +trainer/policy/normal/log_std Mean 1.02378 +trainer/policy/normal/log_std Std 0.247919 +trainer/policy/normal/log_std Max 1.92567 +trainer/policy/normal/log_std Min -0.678435 +eval/num steps total 128558 +eval/num paths total 129 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.212481 +eval/Actions Std 0.842153 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85854 +time/logging (s) 0.00408533 +time/sampling batch (s) 0.293032 +time/saving (s) 0.00438702 +time/training (s) 6.62052 +time/epoch (s) 9.78057 +time/total (s) 1360.53 +Epoch -872 +---------------------------------- --------------- +2022-05-10 13:33:27.347400 PDT | [0] Epoch -871 finished +---------------------------------- --------------- +epoch -871 +replay_buffer/size 999033 +trainer/num train calls 130000 +trainer/Policy Loss -19.3712 +trainer/Log Pis Mean 25.1639 +trainer/Log Pis Std 13.7237 +trainer/Log Pis Max 70.7889 +trainer/Log Pis Min -8.96623 +trainer/policy/mean Mean -0.0310609 +trainer/policy/mean Std 0.90623 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.93684 +trainer/policy/normal/std Std 0.575298 +trainer/policy/normal/std Max 5.00053 +trainer/policy/normal/std Min 0.470212 +trainer/policy/normal/log_std Mean 1.05098 +trainer/policy/normal/log_std Std 0.254615 +trainer/policy/normal/log_std Max 1.60954 +trainer/policy/normal/log_std Min -0.754571 +eval/num steps total 129558 +eval/num paths total 130 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.380147 +eval/Actions Std 0.892283 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69884 +time/logging (s) 0.00371419 +time/sampling batch (s) 0.537176 +time/saving (s) 0.00338673 +time/training (s) 8.06911 +time/epoch (s) 11.3122 +time/total (s) 1371.85 +Epoch -871 +---------------------------------- --------------- +2022-05-10 13:33:36.823160 PDT | [0] Epoch -870 finished +---------------------------------- --------------- +epoch -870 +replay_buffer/size 999033 +trainer/num train calls 131000 +trainer/Policy Loss -19.3544 +trainer/Log Pis Mean 24.6109 +trainer/Log Pis Std 13.5221 +trainer/Log Pis Max 66.8854 +trainer/Log Pis Min -7.77607 +trainer/policy/mean Mean -0.0375805 +trainer/policy/mean Std 0.905998 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84822 +trainer/policy/normal/std Std 0.544844 +trainer/policy/normal/std Max 5.30475 +trainer/policy/normal/std Min 0.441534 +trainer/policy/normal/log_std Mean 1.02229 +trainer/policy/normal/log_std Std 0.242416 +trainer/policy/normal/log_std Max 1.6686 +trainer/policy/normal/log_std Min -0.8175 +eval/num steps total 130558 +eval/num paths total 131 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.096686 +eval/Actions Std 0.907575 +eval/Actions Max 0.999993 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3718 +time/logging (s) 0.00453298 +time/sampling batch (s) 0.280984 +time/saving (s) 0.00353402 +time/training (s) 6.79667 +time/epoch (s) 9.45752 +time/total (s) 1381.31 +Epoch -870 +---------------------------------- --------------- +2022-05-10 13:33:47.305355 PDT | [0] Epoch -869 finished +---------------------------------- --------------- +epoch -869 +replay_buffer/size 999033 +trainer/num train calls 132000 +trainer/Policy Loss -20.5198 +trainer/Log Pis Mean 24.7988 +trainer/Log Pis Std 13.1265 +trainer/Log Pis Max 67.8972 +trainer/Log Pis Min -4.8894 +trainer/policy/mean Mean -0.0570445 +trainer/policy/mean Std 0.901779 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86627 +trainer/policy/normal/std Std 0.576681 +trainer/policy/normal/std Max 6.07126 +trainer/policy/normal/std Min 0.46303 +trainer/policy/normal/log_std Mean 1.0245 +trainer/policy/normal/log_std Std 0.266876 +trainer/policy/normal/log_std Max 1.80357 +trainer/policy/normal/log_std Min -0.769963 +eval/num steps total 131558 +eval/num paths total 132 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0500747 +eval/Actions Std 0.908069 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7199 +time/logging (s) 0.00380258 +time/sampling batch (s) 0.281242 +time/saving (s) 0.00344781 +time/training (s) 7.45317 +time/epoch (s) 10.4616 +time/total (s) 1391.77 +Epoch -869 +---------------------------------- --------------- +2022-05-10 13:33:57.166923 PDT | [0] Epoch -868 finished +---------------------------------- --------------- +epoch -868 +replay_buffer/size 999033 +trainer/num train calls 133000 +trainer/Policy Loss -20.6009 +trainer/Log Pis Mean 24.731 +trainer/Log Pis Std 13.3221 +trainer/Log Pis Max 70.7131 +trainer/Log Pis Min -8.69622 +trainer/policy/mean Mean -0.0263495 +trainer/policy/mean Std 0.90443 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.83779 +trainer/policy/normal/std Std 0.577955 +trainer/policy/normal/std Max 5.16121 +trainer/policy/normal/std Min 0.485355 +trainer/policy/normal/log_std Mean 1.01445 +trainer/policy/normal/log_std Std 0.26473 +trainer/policy/normal/log_std Max 1.64117 +trainer/policy/normal/log_std Min -0.722874 +eval/num steps total 132558 +eval/num paths total 133 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.334592 +eval/Actions Std 0.878052 +eval/Actions Max 0.999984 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60397 +time/logging (s) 0.00378456 +time/sampling batch (s) 0.284073 +time/saving (s) 0.00347701 +time/training (s) 6.94699 +time/epoch (s) 9.8423 +time/total (s) 1401.62 +Epoch -868 +---------------------------------- --------------- +2022-05-10 13:34:07.738026 PDT | [0] Epoch -867 finished +---------------------------------- --------------- +epoch -867 +replay_buffer/size 999033 +trainer/num train calls 134000 +trainer/Policy Loss -20.1138 +trainer/Log Pis Mean 24.9683 +trainer/Log Pis Std 13.2791 +trainer/Log Pis Max 66.8796 +trainer/Log Pis Min -8.43694 +trainer/policy/mean Mean -0.0370589 +trainer/policy/mean Std 0.907731 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87925 +trainer/policy/normal/std Std 0.546926 +trainer/policy/normal/std Max 5.68638 +trainer/policy/normal/std Min 0.412295 +trainer/policy/normal/log_std Mean 1.0325 +trainer/policy/normal/log_std Std 0.249145 +trainer/policy/normal/log_std Max 1.73807 +trainer/policy/normal/log_std Min -0.886016 +eval/num steps total 133154 +eval/num paths total 134 +eval/path length Mean 596 +eval/path length Std 0 +eval/path length Max 596 +eval/path length Min 596 +eval/Rewards Mean 0.00167785 +eval/Rewards Std 0.0409272 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0217021 +eval/Actions Std 0.898159 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.68815 +time/logging (s) 0.00313279 +time/sampling batch (s) 0.281774 +time/saving (s) 0.00416872 +time/training (s) 7.57398 +time/epoch (s) 10.5512 +time/total (s) 1412.17 +Epoch -867 +---------------------------------- --------------- +2022-05-10 13:34:18.051159 PDT | [0] Epoch -866 finished +---------------------------------- --------------- +epoch -866 +replay_buffer/size 999033 +trainer/num train calls 135000 +trainer/Policy Loss -19.0398 +trainer/Log Pis Mean 24.1925 +trainer/Log Pis Std 13.5039 +trainer/Log Pis Max 77.1939 +trainer/Log Pis Min -6.81482 +trainer/policy/mean Mean -0.0379585 +trainer/policy/mean Std 0.90521 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81951 +trainer/policy/normal/std Std 0.560947 +trainer/policy/normal/std Max 5.07793 +trainer/policy/normal/std Min 0.478805 +trainer/policy/normal/log_std Mean 1.00929 +trainer/policy/normal/log_std Std 0.258397 +trainer/policy/normal/log_std Max 1.6249 +trainer/policy/normal/log_std Min -0.736463 +eval/num steps total 134154 +eval/num paths total 135 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.201725 +eval/Actions Std 0.897776 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7835 +time/logging (s) 0.00376347 +time/sampling batch (s) 0.782386 +time/saving (s) 0.00380176 +time/training (s) 6.72081 +time/epoch (s) 10.2943 +time/total (s) 1422.47 +Epoch -866 +---------------------------------- --------------- +2022-05-10 13:34:29.015440 PDT | [0] Epoch -865 finished +---------------------------------- --------------- +epoch -865 +replay_buffer/size 999033 +trainer/num train calls 136000 +trainer/Policy Loss -18.6512 +trainer/Log Pis Mean 24.1355 +trainer/Log Pis Std 13.319 +trainer/Log Pis Max 67.214 +trainer/Log Pis Min -9.75816 +trainer/policy/mean Mean -0.0360721 +trainer/policy/mean Std 0.901827 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.89693 +trainer/policy/normal/std Std 0.564224 +trainer/policy/normal/std Max 5.68191 +trainer/policy/normal/std Min 0.484906 +trainer/policy/normal/log_std Mean 1.03883 +trainer/policy/normal/log_std Std 0.243781 +trainer/policy/normal/log_std Max 1.73729 +trainer/policy/normal/log_std Min -0.7238 +eval/num steps total 135154 +eval/num paths total 136 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.45017 +eval/Actions Std 0.87129 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68496 +time/logging (s) 0.00373142 +time/sampling batch (s) 0.280437 +time/saving (s) 0.00351104 +time/training (s) 7.97232 +time/epoch (s) 10.945 +time/total (s) 1433.42 +Epoch -865 +---------------------------------- --------------- +2022-05-10 13:34:39.819117 PDT | [0] Epoch -864 finished +---------------------------------- --------------- +epoch -864 +replay_buffer/size 999033 +trainer/num train calls 137000 +trainer/Policy Loss -19.2637 +trainer/Log Pis Mean 24.6167 +trainer/Log Pis Std 13.2343 +trainer/Log Pis Max 64.8758 +trainer/Log Pis Min -10.9637 +trainer/policy/mean Mean -0.0329389 +trainer/policy/mean Std 0.908254 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88792 +trainer/policy/normal/std Std 0.575016 +trainer/policy/normal/std Max 5.8287 +trainer/policy/normal/std Min 0.478669 +trainer/policy/normal/log_std Mean 1.03358 +trainer/policy/normal/log_std Std 0.257097 +trainer/policy/normal/log_std Max 1.76279 +trainer/policy/normal/log_std Min -0.736746 +eval/num steps total 136154 +eval/num paths total 137 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.233069 +eval/Actions Std 0.888525 +eval/Actions Max 0.999991 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.99705 +time/logging (s) 0.00372282 +time/sampling batch (s) 0.28057 +time/saving (s) 0.00356175 +time/training (s) 7.49979 +time/epoch (s) 10.7847 +time/total (s) 1444.21 +Epoch -864 +---------------------------------- --------------- +2022-05-10 13:34:50.667848 PDT | [0] Epoch -863 finished +---------------------------------- --------------- +epoch -863 +replay_buffer/size 999033 +trainer/num train calls 138000 +trainer/Policy Loss -20.7177 +trainer/Log Pis Mean 24.7222 +trainer/Log Pis Std 13.2993 +trainer/Log Pis Max 72.0857 +trainer/Log Pis Min -9.75048 +trainer/policy/mean Mean -0.0233864 +trainer/policy/mean Std 0.901019 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86254 +trainer/policy/normal/std Std 0.575825 +trainer/policy/normal/std Max 5.13303 +trainer/policy/normal/std Min 0.431864 +trainer/policy/normal/log_std Mean 1.02385 +trainer/policy/normal/log_std Std 0.261799 +trainer/policy/normal/log_std Max 1.6357 +trainer/policy/normal/log_std Min -0.839644 +eval/num steps total 137154 +eval/num paths total 138 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0651593 +eval/Actions Std 0.907161 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66585 +time/logging (s) 0.00394606 +time/sampling batch (s) 0.530808 +time/saving (s) 0.00354304 +time/training (s) 7.62579 +time/epoch (s) 10.8299 +time/total (s) 1455.04 +Epoch -863 +---------------------------------- --------------- +2022-05-10 13:35:00.563699 PDT | [0] Epoch -862 finished +---------------------------------- --------------- +epoch -862 +replay_buffer/size 999033 +trainer/num train calls 139000 +trainer/Policy Loss -18.8969 +trainer/Log Pis Mean 24.2636 +trainer/Log Pis Std 13.1621 +trainer/Log Pis Max 64.1801 +trainer/Log Pis Min -4.02568 +trainer/policy/mean Mean -0.0430686 +trainer/policy/mean Std 0.908238 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84148 +trainer/policy/normal/std Std 0.561152 +trainer/policy/normal/std Max 4.81468 +trainer/policy/normal/std Min 0.441909 +trainer/policy/normal/log_std Mean 1.01717 +trainer/policy/normal/log_std Std 0.258873 +trainer/policy/normal/log_std Max 1.57167 +trainer/policy/normal/log_std Min -0.816651 +eval/num steps total 138154 +eval/num paths total 139 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0116534 +eval/Actions Std 0.856431 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56033 +time/logging (s) 0.00418776 +time/sampling batch (s) 0.530925 +time/saving (s) 0.00416853 +time/training (s) 6.77719 +time/epoch (s) 9.8768 +time/total (s) 1464.92 +Epoch -862 +---------------------------------- --------------- +2022-05-10 13:35:11.045472 PDT | [0] Epoch -861 finished +---------------------------------- --------------- +epoch -861 +replay_buffer/size 999033 +trainer/num train calls 140000 +trainer/Policy Loss -20.3703 +trainer/Log Pis Mean 23.9982 +trainer/Log Pis Std 13.6789 +trainer/Log Pis Max 67.1281 +trainer/Log Pis Min -12.1101 +trainer/policy/mean Mean -0.0396533 +trainer/policy/mean Std 0.910911 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84324 +trainer/policy/normal/std Std 0.553199 +trainer/policy/normal/std Max 4.54772 +trainer/policy/normal/std Min 0.436225 +trainer/policy/normal/log_std Mean 1.01819 +trainer/policy/normal/log_std Std 0.257789 +trainer/policy/normal/log_std Max 1.51463 +trainer/policy/normal/log_std Min -0.829598 +eval/num steps total 139154 +eval/num paths total 140 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138532 +eval/Actions Std 0.87643 +eval/Actions Max 0.999997 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67801 +time/logging (s) 0.00384721 +time/sampling batch (s) 0.533501 +time/saving (s) 0.00374014 +time/training (s) 7.24259 +time/epoch (s) 10.4617 +time/total (s) 1475.38 +Epoch -861 +---------------------------------- --------------- +2022-05-10 13:35:21.724400 PDT | [0] Epoch -860 finished +---------------------------------- --------------- +epoch -860 +replay_buffer/size 999033 +trainer/num train calls 141000 +trainer/Policy Loss -19.1072 +trainer/Log Pis Mean 24.0524 +trainer/Log Pis Std 12.4787 +trainer/Log Pis Max 63.6015 +trainer/Log Pis Min -6.17011 +trainer/policy/mean Mean -0.0434193 +trainer/policy/mean Std 0.903371 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82876 +trainer/policy/normal/std Std 0.567442 +trainer/policy/normal/std Max 5.75274 +trainer/policy/normal/std Min 0.400833 +trainer/policy/normal/log_std Mean 1.01177 +trainer/policy/normal/log_std Std 0.263243 +trainer/policy/normal/log_std Max 1.74968 +trainer/policy/normal/log_std Min -0.914211 +eval/num steps total 140154 +eval/num paths total 141 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0568167 +eval/Actions Std 0.905977 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68254 +time/logging (s) 0.0037728 +time/sampling batch (s) 0.285189 +time/saving (s) 0.00346505 +time/training (s) 7.68442 +time/epoch (s) 10.6594 +time/total (s) 1486.04 +Epoch -860 +---------------------------------- --------------- +2022-05-10 13:35:31.287526 PDT | [0] Epoch -859 finished +---------------------------------- --------------- +epoch -859 +replay_buffer/size 999033 +trainer/num train calls 142000 +trainer/Policy Loss -20.0101 +trainer/Log Pis Mean 24.0103 +trainer/Log Pis Std 12.8907 +trainer/Log Pis Max 81.0453 +trainer/Log Pis Min -12.1053 +trainer/policy/mean Mean -0.0582814 +trainer/policy/mean Std 0.902796 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.88765 +trainer/policy/normal/std Std 0.551671 +trainer/policy/normal/std Max 5.05161 +trainer/policy/normal/std Min 0.472819 +trainer/policy/normal/log_std Mean 1.03612 +trainer/policy/normal/log_std Std 0.24221 +trainer/policy/normal/log_std Max 1.61971 +trainer/policy/normal/log_std Min -0.749043 +eval/num steps total 141154 +eval/num paths total 142 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0864403 +eval/Actions Std 0.89833 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84189 +time/logging (s) 0.00369443 +time/sampling batch (s) 0.282396 +time/saving (s) 0.00340331 +time/training (s) 6.41239 +time/epoch (s) 9.54377 +time/total (s) 1495.59 +Epoch -859 +---------------------------------- --------------- +2022-05-10 13:35:41.694074 PDT | [0] Epoch -858 finished +---------------------------------- --------------- +epoch -858 +replay_buffer/size 999033 +trainer/num train calls 143000 +trainer/Policy Loss -20.0517 +trainer/Log Pis Mean 25.2171 +trainer/Log Pis Std 14.0193 +trainer/Log Pis Max 69.266 +trainer/Log Pis Min -8.30869 +trainer/policy/mean Mean -0.0392763 +trainer/policy/mean Std 0.904485 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.8697 +trainer/policy/normal/std Std 0.617581 +trainer/policy/normal/std Max 5.78615 +trainer/policy/normal/std Min 0.404232 +trainer/policy/normal/log_std Mean 1.02076 +trainer/policy/normal/log_std Std 0.290664 +trainer/policy/normal/log_std Max 1.75547 +trainer/policy/normal/log_std Min -0.905767 +eval/num steps total 142154 +eval/num paths total 143 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00363139 +eval/Actions Std 0.948555 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75517 +time/logging (s) 0.00374646 +time/sampling batch (s) 0.280221 +time/saving (s) 0.00344342 +time/training (s) 7.34497 +time/epoch (s) 10.3876 +time/total (s) 1505.98 +Epoch -858 +---------------------------------- --------------- +2022-05-10 13:35:52.133424 PDT | [0] Epoch -857 finished +---------------------------------- --------------- +epoch -857 +replay_buffer/size 999033 +trainer/num train calls 144000 +trainer/Policy Loss -19.2252 +trainer/Log Pis Mean 25.1609 +trainer/Log Pis Std 13.0782 +trainer/Log Pis Max 71.7328 +trainer/Log Pis Min -10.4126 +trainer/policy/mean Mean -0.0375203 +trainer/policy/mean Std 0.901759 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83733 +trainer/policy/normal/std Std 0.586693 +trainer/policy/normal/std Max 5.88143 +trainer/policy/normal/std Min 0.391142 +trainer/policy/normal/log_std Mean 1.01307 +trainer/policy/normal/log_std Std 0.271581 +trainer/policy/normal/log_std Max 1.7718 +trainer/policy/normal/log_std Min -0.938684 +eval/num steps total 143154 +eval/num paths total 144 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0115061 +eval/Actions Std 0.890874 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70398 +time/logging (s) 0.00388675 +time/sampling batch (s) 0.281098 +time/saving (s) 0.00346933 +time/training (s) 7.42771 +time/epoch (s) 10.4201 +time/total (s) 1516.4 +Epoch -857 +---------------------------------- --------------- +2022-05-10 13:36:02.235047 PDT | [0] Epoch -856 finished +---------------------------------- --------------- +epoch -856 +replay_buffer/size 999033 +trainer/num train calls 145000 +trainer/Policy Loss -18.726 +trainer/Log Pis Mean 24.5691 +trainer/Log Pis Std 12.7488 +trainer/Log Pis Max 82.3402 +trainer/Log Pis Min -5.26469 +trainer/policy/mean Mean -0.0500576 +trainer/policy/mean Std 0.903836 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85906 +trainer/policy/normal/std Std 0.550461 +trainer/policy/normal/std Max 5.61494 +trainer/policy/normal/std Min 0.43437 +trainer/policy/normal/log_std Mean 1.02527 +trainer/policy/normal/log_std Std 0.248243 +trainer/policy/normal/log_std Max 1.72543 +trainer/policy/normal/log_std Min -0.833859 +eval/num steps total 144154 +eval/num paths total 145 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138397 +eval/Actions Std 0.87508 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67839 +time/logging (s) 0.00422362 +time/sampling batch (s) 0.545241 +time/saving (s) 0.00432432 +time/training (s) 6.84935 +time/epoch (s) 10.0815 +time/total (s) 1526.49 +Epoch -856 +---------------------------------- --------------- +2022-05-10 13:36:12.320952 PDT | [0] Epoch -855 finished +---------------------------------- --------------- +epoch -855 +replay_buffer/size 999033 +trainer/num train calls 146000 +trainer/Policy Loss -18.8674 +trainer/Log Pis Mean 24.2726 +trainer/Log Pis Std 13.3773 +trainer/Log Pis Max 64.8357 +trainer/Log Pis Min -6.72598 +trainer/policy/mean Mean -0.060769 +trainer/policy/mean Std 0.907636 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.93139 +trainer/policy/normal/std Std 0.617946 +trainer/policy/normal/std Max 6.59652 +trainer/policy/normal/std Min 0.398886 +trainer/policy/normal/log_std Mean 1.04406 +trainer/policy/normal/log_std Std 0.281043 +trainer/policy/normal/log_std Max 1.88654 +trainer/policy/normal/log_std Min -0.91908 +eval/num steps total 145154 +eval/num paths total 146 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0729884 +eval/Actions Std 0.891922 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74877 +time/logging (s) 0.00372544 +time/sampling batch (s) 0.531316 +time/saving (s) 0.0034784 +time/training (s) 6.77837 +time/epoch (s) 10.0657 +time/total (s) 1536.56 +Epoch -855 +---------------------------------- --------------- +2022-05-10 13:36:22.006276 PDT | [0] Epoch -854 finished +---------------------------------- --------------- +epoch -854 +replay_buffer/size 999033 +trainer/num train calls 147000 +trainer/Policy Loss -19.543 +trainer/Log Pis Mean 25.4897 +trainer/Log Pis Std 13.4191 +trainer/Log Pis Max 74.3116 +trainer/Log Pis Min -6.24295 +trainer/policy/mean Mean -0.0347144 +trainer/policy/mean Std 0.903353 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89254 +trainer/policy/normal/std Std 0.592005 +trainer/policy/normal/std Max 5.61884 +trainer/policy/normal/std Min 0.472088 +trainer/policy/normal/log_std Mean 1.03263 +trainer/policy/normal/log_std Std 0.271528 +trainer/policy/normal/log_std Max 1.72613 +trainer/policy/normal/log_std Min -0.75059 +eval/num steps total 146154 +eval/num paths total 147 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.215103 +eval/Actions Std 0.88592 +eval/Actions Max 0.999989 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49078 +time/logging (s) 0.00365732 +time/sampling batch (s) 0.531743 +time/saving (s) 0.00345784 +time/training (s) 6.63615 +time/epoch (s) 9.66579 +time/total (s) 1546.23 +Epoch -854 +---------------------------------- --------------- +2022-05-10 13:36:32.851528 PDT | [0] Epoch -853 finished +---------------------------------- --------------- +epoch -853 +replay_buffer/size 999033 +trainer/num train calls 148000 +trainer/Policy Loss -21.0974 +trainer/Log Pis Mean 24.3678 +trainer/Log Pis Std 13.89 +trainer/Log Pis Max 64.9977 +trainer/Log Pis Min -7.45249 +trainer/policy/mean Mean -0.0397982 +trainer/policy/mean Std 0.905902 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.87011 +trainer/policy/normal/std Std 0.559279 +trainer/policy/normal/std Max 5.20403 +trainer/policy/normal/std Min 0.449508 +trainer/policy/normal/log_std Mean 1.02774 +trainer/policy/normal/log_std Std 0.257529 +trainer/policy/normal/log_std Max 1.64943 +trainer/policy/normal/log_std Min -0.799601 +eval/num steps total 147154 +eval/num paths total 148 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.108574 +eval/Actions Std 0.910642 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73159 +time/logging (s) 0.00369916 +time/sampling batch (s) 0.289806 +time/saving (s) 0.00364046 +time/training (s) 7.79687 +time/epoch (s) 10.8256 +time/total (s) 1557.05 +Epoch -853 +---------------------------------- --------------- +2022-05-10 13:36:42.365014 PDT | [0] Epoch -852 finished +---------------------------------- --------------- +epoch -852 +replay_buffer/size 999033 +trainer/num train calls 149000 +trainer/Policy Loss -18.9751 +trainer/Log Pis Mean 23.7787 +trainer/Log Pis Std 13.0171 +trainer/Log Pis Max 59.6428 +trainer/Log Pis Min -7.98864 +trainer/policy/mean Mean -0.0286907 +trainer/policy/mean Std 0.904956 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82283 +trainer/policy/normal/std Std 0.559687 +trainer/policy/normal/std Max 5.04495 +trainer/policy/normal/std Min 0.438691 +trainer/policy/normal/log_std Mean 1.01051 +trainer/policy/normal/log_std Std 0.259607 +trainer/policy/normal/log_std Max 1.61839 +trainer/policy/normal/log_std Min -0.823961 +eval/num steps total 148154 +eval/num paths total 149 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0930597 +eval/Actions Std 0.910188 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68848 +time/logging (s) 0.00391437 +time/sampling batch (s) 0.288024 +time/saving (s) 0.00350315 +time/training (s) 6.51005 +time/epoch (s) 9.49397 +time/total (s) 1566.55 +Epoch -852 +---------------------------------- --------------- +2022-05-10 13:36:51.947059 PDT | [0] Epoch -851 finished +---------------------------------- --------------- +epoch -851 +replay_buffer/size 999033 +trainer/num train calls 150000 +trainer/Policy Loss -19.167 +trainer/Log Pis Mean 23.8564 +trainer/Log Pis Std 12.4134 +trainer/Log Pis Max 67.3464 +trainer/Log Pis Min -2.80813 +trainer/policy/mean Mean -0.0347089 +trainer/policy/mean Std 0.904938 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83739 +trainer/policy/normal/std Std 0.561955 +trainer/policy/normal/std Max 5.17981 +trainer/policy/normal/std Min 0.455361 +trainer/policy/normal/log_std Mean 1.01619 +trainer/policy/normal/log_std Std 0.255713 +trainer/policy/normal/log_std Max 1.64477 +trainer/policy/normal/log_std Min -0.786664 +eval/num steps total 149154 +eval/num paths total 150 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.286292 +eval/Actions Std 0.908969 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67603 +time/logging (s) 0.00427519 +time/sampling batch (s) 0.28836 +time/saving (s) 0.00422581 +time/training (s) 6.58956 +time/epoch (s) 9.56246 +time/total (s) 1576.12 +Epoch -851 +---------------------------------- --------------- +2022-05-10 13:37:01.887457 PDT | [0] Epoch -850 finished +---------------------------------- --------------- +epoch -850 +replay_buffer/size 999033 +trainer/num train calls 151000 +trainer/Policy Loss -19.8067 +trainer/Log Pis Mean 24.1224 +trainer/Log Pis Std 13.6525 +trainer/Log Pis Max 66.4112 +trainer/Log Pis Min -6.70497 +trainer/policy/mean Mean -0.0369708 +trainer/policy/mean Std 0.905725 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85945 +trainer/policy/normal/std Std 0.586796 +trainer/policy/normal/std Max 5.52961 +trainer/policy/normal/std Min 0.456051 +trainer/policy/normal/log_std Mean 1.02074 +trainer/policy/normal/log_std Std 0.273484 +trainer/policy/normal/log_std Max 1.71012 +trainer/policy/normal/log_std Min -0.78515 +eval/num steps total 150154 +eval/num paths total 151 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.223611 +eval/Actions Std 0.874907 +eval/Actions Max 0.999995 +eval/Actions Min -0.999968 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81512 +time/logging (s) 0.0037574 +time/sampling batch (s) 0.532151 +time/saving (s) 0.00365232 +time/training (s) 6.56548 +time/epoch (s) 9.92016 +time/total (s) 1586.04 +Epoch -850 +---------------------------------- --------------- +2022-05-10 13:37:12.187512 PDT | [0] Epoch -849 finished +---------------------------------- --------------- +epoch -849 +replay_buffer/size 999033 +trainer/num train calls 152000 +trainer/Policy Loss -20.2393 +trainer/Log Pis Mean 23.3369 +trainer/Log Pis Std 12.8842 +trainer/Log Pis Max 60.9876 +trainer/Log Pis Min -9.70523 +trainer/policy/mean Mean -0.0435089 +trainer/policy/mean Std 0.901725 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80186 +trainer/policy/normal/std Std 0.559461 +trainer/policy/normal/std Max 6.60883 +trainer/policy/normal/std Min 0.444239 +trainer/policy/normal/log_std Mean 1.00312 +trainer/policy/normal/log_std Std 0.257657 +trainer/policy/normal/log_std Max 1.88841 +trainer/policy/normal/log_std Min -0.811392 +eval/num steps total 151154 +eval/num paths total 152 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0557696 +eval/Actions Std 0.894364 +eval/Actions Max 0.999999 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6032 +time/logging (s) 0.00366622 +time/sampling batch (s) 0.282276 +time/saving (s) 0.00347548 +time/training (s) 7.38791 +time/epoch (s) 10.2805 +time/total (s) 1596.32 +Epoch -849 +---------------------------------- --------------- +2022-05-10 13:37:22.749789 PDT | [0] Epoch -848 finished +---------------------------------- --------------- +epoch -848 +replay_buffer/size 999033 +trainer/num train calls 153000 +trainer/Policy Loss -19.5789 +trainer/Log Pis Mean 23.7651 +trainer/Log Pis Std 13.266 +trainer/Log Pis Max 64.9377 +trainer/Log Pis Min -7.99235 +trainer/policy/mean Mean -0.0405487 +trainer/policy/mean Std 0.901316 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.90292 +trainer/policy/normal/std Std 0.584256 +trainer/policy/normal/std Max 5.56924 +trainer/policy/normal/std Min 0.436217 +trainer/policy/normal/log_std Mean 1.03784 +trainer/policy/normal/log_std Std 0.262298 +trainer/policy/normal/log_std Max 1.71726 +trainer/policy/normal/log_std Min -0.829615 +eval/num steps total 152154 +eval/num paths total 153 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0694443 +eval/Actions Std 0.886089 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64046 +time/logging (s) 0.00368295 +time/sampling batch (s) 0.283698 +time/saving (s) 0.00349237 +time/training (s) 7.61157 +time/epoch (s) 10.5429 +time/total (s) 1606.87 +Epoch -848 +---------------------------------- --------------- +2022-05-10 13:37:32.232972 PDT | [0] Epoch -847 finished +---------------------------------- --------------- +epoch -847 +replay_buffer/size 999033 +trainer/num train calls 154000 +trainer/Policy Loss -19.346 +trainer/Log Pis Mean 23.9633 +trainer/Log Pis Std 13.8224 +trainer/Log Pis Max 77.832 +trainer/Log Pis Min -16.5132 +trainer/policy/mean Mean -0.0213512 +trainer/policy/mean Std 0.906387 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8776 +trainer/policy/normal/std Std 0.57317 +trainer/policy/normal/std Max 5.3263 +trainer/policy/normal/std Min 0.457954 +trainer/policy/normal/log_std Mean 1.02981 +trainer/policy/normal/log_std Std 0.257864 +trainer/policy/normal/log_std Max 1.67266 +trainer/policy/normal/log_std Min -0.780986 +eval/num steps total 153154 +eval/num paths total 154 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0494791 +eval/Actions Std 0.896104 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6819 +time/logging (s) 0.00375056 +time/sampling batch (s) 0.283391 +time/saving (s) 0.00355236 +time/training (s) 6.49138 +time/epoch (s) 9.46397 +time/total (s) 1616.34 +Epoch -847 +---------------------------------- --------------- +2022-05-10 13:37:42.507826 PDT | [0] Epoch -846 finished +---------------------------------- --------------- +epoch -846 +replay_buffer/size 999033 +trainer/num train calls 155000 +trainer/Policy Loss -19.5417 +trainer/Log Pis Mean 24.7315 +trainer/Log Pis Std 13.1265 +trainer/Log Pis Max 77.583 +trainer/Log Pis Min -5.68957 +trainer/policy/mean Mean -0.0110053 +trainer/policy/mean Std 0.90882 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87448 +trainer/policy/normal/std Std 0.574771 +trainer/policy/normal/std Max 6.84571 +trainer/policy/normal/std Min 0.431443 +trainer/policy/normal/log_std Mean 1.02905 +trainer/policy/normal/log_std Std 0.255189 +trainer/policy/normal/log_std Max 1.92362 +trainer/policy/normal/log_std Min -0.840619 +eval/num steps total 154154 +eval/num paths total 155 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0656703 +eval/Actions Std 0.896984 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48746 +time/logging (s) 0.00368277 +time/sampling batch (s) 0.286137 +time/saving (s) 0.00349801 +time/training (s) 7.47427 +time/epoch (s) 10.2551 +time/total (s) 1626.59 +Epoch -846 +---------------------------------- --------------- +2022-05-10 13:37:53.369251 PDT | [0] Epoch -845 finished +---------------------------------- --------------- +epoch -845 +replay_buffer/size 999033 +trainer/num train calls 156000 +trainer/Policy Loss -19.758 +trainer/Log Pis Mean 24.5632 +trainer/Log Pis Std 13.5432 +trainer/Log Pis Max 73.3408 +trainer/Log Pis Min -8.70411 +trainer/policy/mean Mean -0.0272792 +trainer/policy/mean Std 0.904766 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8728 +trainer/policy/normal/std Std 0.579003 +trainer/policy/normal/std Max 5.62493 +trainer/policy/normal/std Min 0.382278 +trainer/policy/normal/log_std Mean 1.02768 +trainer/policy/normal/log_std Std 0.25905 +trainer/policy/normal/log_std Max 1.72721 +trainer/policy/normal/log_std Min -0.961606 +eval/num steps total 155154 +eval/num paths total 156 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00856137 +eval/Actions Std 0.906975 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48262 +time/logging (s) 0.00409837 +time/sampling batch (s) 0.530127 +time/saving (s) 0.00416244 +time/training (s) 7.82148 +time/epoch (s) 10.8425 +time/total (s) 1637.44 +Epoch -845 +---------------------------------- --------------- +2022-05-10 13:38:03.418965 PDT | [0] Epoch -844 finished +---------------------------------- --------------- +epoch -844 +replay_buffer/size 999033 +trainer/num train calls 157000 +trainer/Policy Loss -19.7419 +trainer/Log Pis Mean 24.6913 +trainer/Log Pis Std 13.4192 +trainer/Log Pis Max 88.8332 +trainer/Log Pis Min -4.68881 +trainer/policy/mean Mean -0.0552328 +trainer/policy/mean Std 0.901241 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.86657 +trainer/policy/normal/std Std 0.575472 +trainer/policy/normal/std Max 5.68941 +trainer/policy/normal/std Min 0.393193 +trainer/policy/normal/log_std Mean 1.02555 +trainer/policy/normal/log_std Std 0.26024 +trainer/policy/normal/log_std Max 1.73861 +trainer/policy/normal/log_std Min -0.933455 +eval/num steps total 156154 +eval/num paths total 157 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0131606 +eval/Actions Std 0.930252 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61907 +time/logging (s) 0.00377316 +time/sampling batch (s) 0.280427 +time/saving (s) 0.00355896 +time/training (s) 7.12288 +time/epoch (s) 10.0297 +time/total (s) 1647.47 +Epoch -844 +---------------------------------- --------------- +2022-05-10 13:38:14.065023 PDT | [0] Epoch -843 finished +---------------------------------- --------------- +epoch -843 +replay_buffer/size 999033 +trainer/num train calls 158000 +trainer/Policy Loss -20.4058 +trainer/Log Pis Mean 25.4211 +trainer/Log Pis Std 13.6707 +trainer/Log Pis Max 72.4363 +trainer/Log Pis Min -8.05296 +trainer/policy/mean Mean -0.0434537 +trainer/policy/mean Std 0.903456 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85312 +trainer/policy/normal/std Std 0.569392 +trainer/policy/normal/std Max 4.85179 +trainer/policy/normal/std Min 0.453533 +trainer/policy/normal/log_std Mean 1.02023 +trainer/policy/normal/log_std Std 0.265396 +trainer/policy/normal/log_std Max 1.57935 +trainer/policy/normal/log_std Min -0.790686 +eval/num steps total 157154 +eval/num paths total 158 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.250933 +eval/Actions Std 0.8676 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58198 +time/logging (s) 0.00382229 +time/sampling batch (s) 0.278736 +time/saving (s) 0.00346481 +time/training (s) 7.75894 +time/epoch (s) 10.627 +time/total (s) 1658.1 +Epoch -843 +---------------------------------- --------------- +2022-05-10 13:38:24.276735 PDT | [0] Epoch -842 finished +---------------------------------- --------------- +epoch -842 +replay_buffer/size 999033 +trainer/num train calls 159000 +trainer/Policy Loss -18.9196 +trainer/Log Pis Mean 24.5297 +trainer/Log Pis Std 13.4067 +trainer/Log Pis Max 66.274 +trainer/Log Pis Min -7.08227 +trainer/policy/mean Mean -0.0399508 +trainer/policy/mean Std 0.903074 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84023 +trainer/policy/normal/std Std 0.586993 +trainer/policy/normal/std Max 5.1527 +trainer/policy/normal/std Min 0.453214 +trainer/policy/normal/log_std Mean 1.01316 +trainer/policy/normal/log_std Std 0.278227 +trainer/policy/normal/log_std Max 1.63952 +trainer/policy/normal/log_std Min -0.791391 +eval/num steps total 158154 +eval/num paths total 159 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00933806 +eval/Actions Std 0.919418 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63755 +time/logging (s) 0.00373231 +time/sampling batch (s) 0.281643 +time/saving (s) 0.00346069 +time/training (s) 7.26592 +time/epoch (s) 10.1923 +time/total (s) 1668.3 +Epoch -842 +---------------------------------- --------------- +2022-05-10 13:38:34.683057 PDT | [0] Epoch -841 finished +---------------------------------- --------------- +epoch -841 +replay_buffer/size 999033 +trainer/num train calls 160000 +trainer/Policy Loss -19.6208 +trainer/Log Pis Mean 24.1359 +trainer/Log Pis Std 12.7262 +trainer/Log Pis Max 62.0285 +trainer/Log Pis Min -9.0398 +trainer/policy/mean Mean -0.0377325 +trainer/policy/mean Std 0.904505 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.89197 +trainer/policy/normal/std Std 0.582898 +trainer/policy/normal/std Max 5.15776 +trainer/policy/normal/std Min 0.491848 +trainer/policy/normal/log_std Mean 1.03359 +trainer/policy/normal/log_std Std 0.265346 +trainer/policy/normal/log_std Max 1.6405 +trainer/policy/normal/log_std Min -0.709586 +eval/num steps total 159154 +eval/num paths total 160 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0135673 +eval/Actions Std 0.905921 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57793 +time/logging (s) 0.0037458 +time/sampling batch (s) 0.282534 +time/saving (s) 0.00345733 +time/training (s) 7.51916 +time/epoch (s) 10.3868 +time/total (s) 1678.69 +Epoch -841 +---------------------------------- --------------- +2022-05-10 13:38:45.362662 PDT | [0] Epoch -840 finished +---------------------------------- --------------- +epoch -840 +replay_buffer/size 999033 +trainer/num train calls 161000 +trainer/Policy Loss -18.6905 +trainer/Log Pis Mean 23.9229 +trainer/Log Pis Std 13.322 +trainer/Log Pis Max 69.5988 +trainer/Log Pis Min -8.60245 +trainer/policy/mean Mean -0.0637214 +trainer/policy/mean Std 0.899347 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84192 +trainer/policy/normal/std Std 0.601542 +trainer/policy/normal/std Max 5.50184 +trainer/policy/normal/std Min 0.437733 +trainer/policy/normal/log_std Mean 1.01216 +trainer/policy/normal/log_std Std 0.286453 +trainer/policy/normal/log_std Max 1.70508 +trainer/policy/normal/log_std Min -0.826146 +eval/num steps total 160154 +eval/num paths total 161 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00712541 +eval/Actions Std 0.84945 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66057 +time/logging (s) 0.00400268 +time/sampling batch (s) 0.533243 +time/saving (s) 0.00403163 +time/training (s) 7.45843 +time/epoch (s) 10.6603 +time/total (s) 1689.35 +Epoch -840 +---------------------------------- --------------- +2022-05-10 13:38:55.244023 PDT | [0] Epoch -839 finished +---------------------------------- --------------- +epoch -839 +replay_buffer/size 999033 +trainer/num train calls 162000 +trainer/Policy Loss -19.8112 +trainer/Log Pis Mean 24.797 +trainer/Log Pis Std 13.3717 +trainer/Log Pis Max 74.7985 +trainer/Log Pis Min -4.34769 +trainer/policy/mean Mean -0.0165552 +trainer/policy/mean Std 0.904718 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.93816 +trainer/policy/normal/std Std 0.594402 +trainer/policy/normal/std Max 5.61301 +trainer/policy/normal/std Min 0.438101 +trainer/policy/normal/log_std Mean 1.0505 +trainer/policy/normal/log_std Std 0.256764 +trainer/policy/normal/log_std Max 1.72509 +trainer/policy/normal/log_std Min -0.825306 +eval/num steps total 161154 +eval/num paths total 162 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.23523 +eval/Actions Std 0.583204 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6468 +time/logging (s) 0.00379168 +time/sampling batch (s) 0.281044 +time/saving (s) 0.0036074 +time/training (s) 6.92626 +time/epoch (s) 9.86151 +time/total (s) 1699.21 +Epoch -839 +---------------------------------- --------------- +2022-05-10 13:39:06.322930 PDT | [0] Epoch -838 finished +---------------------------------- --------------- +epoch -838 +replay_buffer/size 999033 +trainer/num train calls 163000 +trainer/Policy Loss -19.4928 +trainer/Log Pis Mean 24.7062 +trainer/Log Pis Std 13.0696 +trainer/Log Pis Max 75.2241 +trainer/Log Pis Min -6.0759 +trainer/policy/mean Mean -0.0382367 +trainer/policy/mean Std 0.905258 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83942 +trainer/policy/normal/std Std 0.584169 +trainer/policy/normal/std Max 6.68867 +trainer/policy/normal/std Min 0.3934 +trainer/policy/normal/log_std Mean 1.01406 +trainer/policy/normal/log_std Std 0.270482 +trainer/policy/normal/log_std Max 1.90041 +trainer/policy/normal/log_std Min -0.932929 +eval/num steps total 162154 +eval/num paths total 163 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0103485 +eval/Actions Std 0.897852 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64217 +time/logging (s) 0.00385053 +time/sampling batch (s) 0.281473 +time/saving (s) 0.00354356 +time/training (s) 8.12851 +time/epoch (s) 11.0596 +time/total (s) 1710.28 +Epoch -838 +---------------------------------- --------------- +2022-05-10 13:39:16.951625 PDT | [0] Epoch -837 finished +---------------------------------- --------------- +epoch -837 +replay_buffer/size 999033 +trainer/num train calls 164000 +trainer/Policy Loss -18.8135 +trainer/Log Pis Mean 23.3646 +trainer/Log Pis Std 12.418 +trainer/Log Pis Max 64.7363 +trainer/Log Pis Min -17.8452 +trainer/policy/mean Mean -0.0191787 +trainer/policy/mean Std 0.907926 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80715 +trainer/policy/normal/std Std 0.592225 +trainer/policy/normal/std Max 6.41252 +trainer/policy/normal/std Min 0.446601 +trainer/policy/normal/log_std Mean 1.001 +trainer/policy/normal/log_std Std 0.277909 +trainer/policy/normal/log_std Max 1.85825 +trainer/policy/normal/log_std Min -0.806089 +eval/num steps total 163154 +eval/num paths total 164 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0557049 +eval/Actions Std 0.904444 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56862 +time/logging (s) 0.0036945 +time/sampling batch (s) 0.287067 +time/saving (s) 0.00338509 +time/training (s) 7.74606 +time/epoch (s) 10.6088 +time/total (s) 1720.89 +Epoch -837 +---------------------------------- --------------- +2022-05-10 13:39:26.285986 PDT | [0] Epoch -836 finished +---------------------------------- --------------- +epoch -836 +replay_buffer/size 999033 +trainer/num train calls 165000 +trainer/Policy Loss -20.0941 +trainer/Log Pis Mean 24.0734 +trainer/Log Pis Std 13.0632 +trainer/Log Pis Max 67.8479 +trainer/Log Pis Min -13.6384 +trainer/policy/mean Mean -0.0555518 +trainer/policy/mean Std 0.903954 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81331 +trainer/policy/normal/std Std 0.591198 +trainer/policy/normal/std Max 6.73503 +trainer/policy/normal/std Min 0.421786 +trainer/policy/normal/log_std Mean 1.00265 +trainer/policy/normal/log_std Std 0.283529 +trainer/policy/normal/log_std Max 1.90732 +trainer/policy/normal/log_std Min -0.863257 +eval/num steps total 164154 +eval/num paths total 165 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139035 +eval/Actions Std 0.939564 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48301 +time/logging (s) 0.00380222 +time/sampling batch (s) 0.286817 +time/saving (s) 0.00353411 +time/training (s) 6.53766 +time/epoch (s) 9.31482 +time/total (s) 1730.21 +Epoch -836 +---------------------------------- --------------- +2022-05-10 13:39:36.183196 PDT | [0] Epoch -835 finished +---------------------------------- --------------- +epoch -835 +replay_buffer/size 999033 +trainer/num train calls 166000 +trainer/Policy Loss -19.5535 +trainer/Log Pis Mean 23.6579 +trainer/Log Pis Std 13.3906 +trainer/Log Pis Max 64.8113 +trainer/Log Pis Min -8.52681 +trainer/policy/mean Mean -0.0324933 +trainer/policy/mean Std 0.909273 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80614 +trainer/policy/normal/std Std 0.556083 +trainer/policy/normal/std Max 5.07494 +trainer/policy/normal/std Min 0.458477 +trainer/policy/normal/log_std Mean 1.00443 +trainer/policy/normal/log_std Std 0.259901 +trainer/policy/normal/log_std Max 1.62431 +trainer/policy/normal/log_std Min -0.779845 +eval/num steps total 165154 +eval/num paths total 166 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0022817 +eval/Actions Std 0.968874 +eval/Actions Max 0.99999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71752 +time/logging (s) 0.00419067 +time/sampling batch (s) 0.286067 +time/saving (s) 0.0040831 +time/training (s) 6.8659 +time/epoch (s) 9.87776 +time/total (s) 1740.09 +Epoch -835 +---------------------------------- --------------- +2022-05-10 13:39:45.350847 PDT | [0] Epoch -834 finished +---------------------------------- --------------- +epoch -834 +replay_buffer/size 999033 +trainer/num train calls 167000 +trainer/Policy Loss -19.7481 +trainer/Log Pis Mean 24.8971 +trainer/Log Pis Std 13.785 +trainer/Log Pis Max 74.7824 +trainer/Log Pis Min -10.6293 +trainer/policy/mean Mean -0.0469628 +trainer/policy/mean Std 0.901553 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.92044 +trainer/policy/normal/std Std 0.611596 +trainer/policy/normal/std Max 5.38956 +trainer/policy/normal/std Min 0.449939 +trainer/policy/normal/log_std Mean 1.04117 +trainer/policy/normal/log_std Std 0.275497 +trainer/policy/normal/log_std Max 1.68446 +trainer/policy/normal/log_std Min -0.798644 +eval/num steps total 166154 +eval/num paths total 167 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0694518 +eval/Actions Std 0.892458 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41673 +time/logging (s) 0.00385455 +time/sampling batch (s) 0.280597 +time/saving (s) 0.00385327 +time/training (s) 6.44274 +time/epoch (s) 9.14777 +time/total (s) 1749.24 +Epoch -834 +---------------------------------- --------------- +2022-05-10 13:39:55.365565 PDT | [0] Epoch -833 finished +---------------------------------- --------------- +epoch -833 +replay_buffer/size 999033 +trainer/num train calls 168000 +trainer/Policy Loss -19.5565 +trainer/Log Pis Mean 23.2153 +trainer/Log Pis Std 13.1056 +trainer/Log Pis Max 63.7378 +trainer/Log Pis Min -11.3471 +trainer/policy/mean Mean -0.0432535 +trainer/policy/mean Std 0.905766 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.85622 +trainer/policy/normal/std Std 0.563802 +trainer/policy/normal/std Max 5.38215 +trainer/policy/normal/std Min 0.424959 +trainer/policy/normal/log_std Mean 1.02237 +trainer/policy/normal/log_std Std 0.259261 +trainer/policy/normal/log_std Max 1.68309 +trainer/policy/normal/log_std Min -0.855763 +eval/num steps total 167154 +eval/num paths total 168 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0724549 +eval/Actions Std 0.930273 +eval/Actions Max 0.999987 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63015 +time/logging (s) 0.00373167 +time/sampling batch (s) 0.282946 +time/saving (s) 0.00348658 +time/training (s) 7.07453 +time/epoch (s) 9.99484 +time/total (s) 1759.24 +Epoch -833 +---------------------------------- --------------- +2022-05-10 13:40:05.032348 PDT | [0] Epoch -832 finished +---------------------------------- --------------- +epoch -832 +replay_buffer/size 999033 +trainer/num train calls 169000 +trainer/Policy Loss -18.5486 +trainer/Log Pis Mean 24.1042 +trainer/Log Pis Std 14.1599 +trainer/Log Pis Max 85.2177 +trainer/Log Pis Min -10.8631 +trainer/policy/mean Mean -0.042604 +trainer/policy/mean Std 0.898219 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87079 +trainer/policy/normal/std Std 0.583538 +trainer/policy/normal/std Max 5.27608 +trainer/policy/normal/std Min 0.45476 +trainer/policy/normal/log_std Mean 1.02576 +trainer/policy/normal/log_std Std 0.267161 +trainer/policy/normal/log_std Max 1.66318 +trainer/policy/normal/log_std Min -0.787985 +eval/num steps total 168154 +eval/num paths total 169 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.226091 +eval/Actions Std 0.924253 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6027 +time/logging (s) 0.00371209 +time/sampling batch (s) 0.280524 +time/saving (s) 0.00347312 +time/training (s) 6.75701 +time/epoch (s) 9.64742 +time/total (s) 1768.89 +Epoch -832 +---------------------------------- --------------- +2022-05-10 13:40:14.630444 PDT | [0] Epoch -831 finished +---------------------------------- --------------- +epoch -831 +replay_buffer/size 999033 +trainer/num train calls 170000 +trainer/Policy Loss -19.1489 +trainer/Log Pis Mean 24 +trainer/Log Pis Std 12.4673 +trainer/Log Pis Max 60.4702 +trainer/Log Pis Min -10.5655 +trainer/policy/mean Mean -0.0281145 +trainer/policy/mean Std 0.903662 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79585 +trainer/policy/normal/std Std 0.610106 +trainer/policy/normal/std Max 6.03753 +trainer/policy/normal/std Min 0.452558 +trainer/policy/normal/log_std Mean 0.994532 +trainer/policy/normal/log_std Std 0.289449 +trainer/policy/normal/log_std Max 1.798 +trainer/policy/normal/log_std Min -0.79284 +eval/num steps total 169154 +eval/num paths total 170 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.376524 +eval/Actions Std 0.862869 +eval/Actions Max 0.999974 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75509 +time/logging (s) 0.00372017 +time/sampling batch (s) 0.278742 +time/saving (s) 0.00351168 +time/training (s) 6.53783 +time/epoch (s) 9.5789 +time/total (s) 1778.47 +Epoch -831 +---------------------------------- --------------- +2022-05-10 13:40:24.657824 PDT | [0] Epoch -830 finished +---------------------------------- --------------- +epoch -830 +replay_buffer/size 999033 +trainer/num train calls 171000 +trainer/Policy Loss -20.022 +trainer/Log Pis Mean 24.6793 +trainer/Log Pis Std 12.5308 +trainer/Log Pis Max 63.2461 +trainer/Log Pis Min -6.10561 +trainer/policy/mean Mean -0.0685093 +trainer/policy/mean Std 0.904182 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83188 +trainer/policy/normal/std Std 0.568979 +trainer/policy/normal/std Max 4.93129 +trainer/policy/normal/std Min 0.462461 +trainer/policy/normal/log_std Mean 1.01211 +trainer/policy/normal/log_std Std 0.268555 +trainer/policy/normal/log_std Max 1.5956 +trainer/policy/normal/log_std Min -0.771193 +eval/num steps total 170154 +eval/num paths total 171 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0908368 +eval/Actions Std 0.913618 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62964 +time/logging (s) 0.00414541 +time/sampling batch (s) 0.278303 +time/saving (s) 0.00408469 +time/training (s) 7.0924 +time/epoch (s) 10.0086 +time/total (s) 1788.48 +Epoch -830 +---------------------------------- --------------- +2022-05-10 13:40:34.770817 PDT | [0] Epoch -829 finished +---------------------------------- --------------- +epoch -829 +replay_buffer/size 999033 +trainer/num train calls 172000 +trainer/Policy Loss -20.0175 +trainer/Log Pis Mean 24.4742 +trainer/Log Pis Std 12.9152 +trainer/Log Pis Max 65.3573 +trainer/Log Pis Min -7.48038 +trainer/policy/mean Mean -0.0578813 +trainer/policy/mean Std 0.904899 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.86248 +trainer/policy/normal/std Std 0.589066 +trainer/policy/normal/std Max 6.49514 +trainer/policy/normal/std Min 0.439346 +trainer/policy/normal/log_std Mean 1.02287 +trainer/policy/normal/log_std Std 0.265193 +trainer/policy/normal/log_std Max 1.87105 +trainer/policy/normal/log_std Min -0.822469 +eval/num steps total 171154 +eval/num paths total 172 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0616566 +eval/Actions Std 0.902024 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.32887 +time/logging (s) 0.00377538 +time/sampling batch (s) 0.533093 +time/saving (s) 0.00383443 +time/training (s) 7.22331 +time/epoch (s) 10.0929 +time/total (s) 1798.57 +Epoch -829 +---------------------------------- --------------- +2022-05-10 13:40:43.872441 PDT | [0] Epoch -828 finished +---------------------------------- --------------- +epoch -828 +replay_buffer/size 999033 +trainer/num train calls 173000 +trainer/Policy Loss -21.0509 +trainer/Log Pis Mean 24.1523 +trainer/Log Pis Std 12.5866 +trainer/Log Pis Max 64.4611 +trainer/Log Pis Min -6.88741 +trainer/policy/mean Mean -0.00990105 +trainer/policy/mean Std 0.909643 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.77335 +trainer/policy/normal/std Std 0.569132 +trainer/policy/normal/std Max 5.51757 +trainer/policy/normal/std Min 0.432618 +trainer/policy/normal/log_std Mean 0.990307 +trainer/policy/normal/log_std Std 0.272327 +trainer/policy/normal/log_std Max 1.70794 +trainer/policy/normal/log_std Min -0.837901 +eval/num steps total 172154 +eval/num paths total 173 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0804938 +eval/Actions Std 0.911665 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72545 +time/logging (s) 0.0036537 +time/sampling batch (s) 0.281694 +time/saving (s) 0.00340409 +time/training (s) 6.06765 +time/epoch (s) 9.08186 +time/total (s) 1807.66 +Epoch -828 +---------------------------------- --------------- +2022-05-10 13:40:54.083666 PDT | [0] Epoch -827 finished +---------------------------------- --------------- +epoch -827 +replay_buffer/size 999033 +trainer/num train calls 174000 +trainer/Policy Loss -19.1418 +trainer/Log Pis Mean 24.1079 +trainer/Log Pis Std 12.7818 +trainer/Log Pis Max 76.8254 +trainer/Log Pis Min -8.88212 +trainer/policy/mean Mean -0.0426762 +trainer/policy/mean Std 0.904083 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82173 +trainer/policy/normal/std Std 0.590265 +trainer/policy/normal/std Max 6.08539 +trainer/policy/normal/std Min 0.423156 +trainer/policy/normal/log_std Mean 1.00687 +trainer/policy/normal/log_std Std 0.275432 +trainer/policy/normal/log_std Max 1.80589 +trainer/policy/normal/log_std Min -0.860014 +eval/num steps total 173154 +eval/num paths total 174 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0480366 +eval/Actions Std 0.909531 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61947 +time/logging (s) 0.00366224 +time/sampling batch (s) 0.283014 +time/saving (s) 0.00345963 +time/training (s) 7.28206 +time/epoch (s) 10.1917 +time/total (s) 1817.85 +Epoch -827 +---------------------------------- --------------- +2022-05-10 13:41:04.803938 PDT | [0] Epoch -826 finished +---------------------------------- --------------- +epoch -826 +replay_buffer/size 999033 +trainer/num train calls 175000 +trainer/Policy Loss -18.5913 +trainer/Log Pis Mean 25.3858 +trainer/Log Pis Std 13.6219 +trainer/Log Pis Max 65.0072 +trainer/Log Pis Min -11.1932 +trainer/policy/mean Mean -0.0703834 +trainer/policy/mean Std 0.904141 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.93697 +trainer/policy/normal/std Std 0.600404 +trainer/policy/normal/std Max 5.87671 +trainer/policy/normal/std Min 0.457211 +trainer/policy/normal/log_std Mean 1.04922 +trainer/policy/normal/log_std Std 0.262298 +trainer/policy/normal/log_std Max 1.771 +trainer/policy/normal/log_std Min -0.78261 +eval/num steps total 174154 +eval/num paths total 175 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.168849 +eval/Actions Std 0.901382 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87688 +time/logging (s) 0.00434532 +time/sampling batch (s) 0.531847 +time/saving (s) 0.00385355 +time/training (s) 7.28455 +time/epoch (s) 10.7015 +time/total (s) 1828.56 +Epoch -826 +---------------------------------- --------------- +2022-05-10 13:41:14.958185 PDT | [0] Epoch -825 finished +---------------------------------- --------------- +epoch -825 +replay_buffer/size 999033 +trainer/num train calls 176000 +trainer/Policy Loss -19.4162 +trainer/Log Pis Mean 24.7582 +trainer/Log Pis Std 13.0734 +trainer/Log Pis Max 65.8654 +trainer/Log Pis Min -11.5625 +trainer/policy/mean Mean -0.0468178 +trainer/policy/mean Std 0.909079 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86192 +trainer/policy/normal/std Std 0.574119 +trainer/policy/normal/std Max 4.85128 +trainer/policy/normal/std Min 0.384791 +trainer/policy/normal/log_std Mean 1.02325 +trainer/policy/normal/log_std Std 0.265305 +trainer/policy/normal/log_std Max 1.57924 +trainer/policy/normal/log_std Min -0.955054 +eval/num steps total 175154 +eval/num paths total 176 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.34438 +eval/Actions Std 0.86821 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65383 +time/logging (s) 0.00366242 +time/sampling batch (s) 0.528105 +time/saving (s) 0.00346309 +time/training (s) 6.94479 +time/epoch (s) 10.1339 +time/total (s) 1838.7 +Epoch -825 +---------------------------------- --------------- +2022-05-10 13:41:24.856503 PDT | [0] Epoch -824 finished +---------------------------------- --------------- +epoch -824 +replay_buffer/size 999033 +trainer/num train calls 177000 +trainer/Policy Loss -19.1952 +trainer/Log Pis Mean 24.3708 +trainer/Log Pis Std 13.0025 +trainer/Log Pis Max 68.934 +trainer/Log Pis Min -4.69639 +trainer/policy/mean Mean -0.0258674 +trainer/policy/mean Std 0.908095 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.7792 +trainer/policy/normal/std Std 0.592969 +trainer/policy/normal/std Max 4.96141 +trainer/policy/normal/std Min 0.41457 +trainer/policy/normal/log_std Mean 0.988303 +trainer/policy/normal/log_std Std 0.294477 +trainer/policy/normal/log_std Max 1.60169 +trainer/policy/normal/log_std Min -0.880513 +eval/num steps total 176154 +eval/num paths total 177 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0882721 +eval/Actions Std 0.923548 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41599 +time/logging (s) 0.00403407 +time/sampling batch (s) 0.282961 +time/saving (s) 0.00411114 +time/training (s) 7.17206 +time/epoch (s) 9.87916 +time/total (s) 1848.58 +Epoch -824 +---------------------------------- --------------- +2022-05-10 13:41:35.213489 PDT | [0] Epoch -823 finished +---------------------------------- --------------- +epoch -823 +replay_buffer/size 999033 +trainer/num train calls 178000 +trainer/Policy Loss -19.2634 +trainer/Log Pis Mean 24.117 +trainer/Log Pis Std 13.0171 +trainer/Log Pis Max 69.8728 +trainer/Log Pis Min -8.29782 +trainer/policy/mean Mean -0.0554253 +trainer/policy/mean Std 0.904309 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84109 +trainer/policy/normal/std Std 0.613571 +trainer/policy/normal/std Max 5.35004 +trainer/policy/normal/std Min 0.356481 +trainer/policy/normal/log_std Mean 1.01022 +trainer/policy/normal/log_std Std 0.293787 +trainer/policy/normal/log_std Max 1.6771 +trainer/policy/normal/log_std Min -1.03147 +eval/num steps total 177154 +eval/num paths total 178 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.07328 +eval/Actions Std 0.915055 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7247 +time/logging (s) 0.00372075 +time/sampling batch (s) 0.530689 +time/saving (s) 0.0035614 +time/training (s) 7.07438 +time/epoch (s) 10.3371 +time/total (s) 1858.92 +Epoch -823 +---------------------------------- --------------- +2022-05-10 13:41:44.824213 PDT | [0] Epoch -822 finished +---------------------------------- --------------- +epoch -822 +replay_buffer/size 999033 +trainer/num train calls 179000 +trainer/Policy Loss -19.5178 +trainer/Log Pis Mean 23.8815 +trainer/Log Pis Std 13.1483 +trainer/Log Pis Max 66.3986 +trainer/Log Pis Min -10.3151 +trainer/policy/mean Mean -0.0396979 +trainer/policy/mean Std 0.901866 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78866 +trainer/policy/normal/std Std 0.588751 +trainer/policy/normal/std Max 5.38557 +trainer/policy/normal/std Min 0.418966 +trainer/policy/normal/log_std Mean 0.994556 +trainer/policy/normal/log_std Std 0.277078 +trainer/policy/normal/log_std Max 1.68372 +trainer/policy/normal/log_std Min -0.869964 +eval/num steps total 178154 +eval/num paths total 179 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.489429 +eval/Actions Std 0.86874 +eval/Actions Max 0.999947 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61529 +time/logging (s) 0.00366178 +time/sampling batch (s) 0.27882 +time/saving (s) 0.00342641 +time/training (s) 6.69002 +time/epoch (s) 9.59122 +time/total (s) 1868.51 +Epoch -822 +---------------------------------- --------------- +2022-05-10 13:41:54.517876 PDT | [0] Epoch -821 finished +---------------------------------- --------------- +epoch -821 +replay_buffer/size 999033 +trainer/num train calls 180000 +trainer/Policy Loss -19.054 +trainer/Log Pis Mean 24.4333 +trainer/Log Pis Std 13.0998 +trainer/Log Pis Max 75.472 +trainer/Log Pis Min -7.9806 +trainer/policy/mean Mean -0.0422265 +trainer/policy/mean Std 0.909111 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.8625 +trainer/policy/normal/std Std 0.601657 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.397143 +trainer/policy/normal/log_std Mean 1.02047 +trainer/policy/normal/log_std Std 0.279832 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.923459 +eval/num steps total 179154 +eval/num paths total 180 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.2995 +eval/Actions Std 0.833138 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.579 +time/logging (s) 0.00369327 +time/sampling batch (s) 0.283097 +time/saving (s) 0.00338503 +time/training (s) 6.80482 +time/epoch (s) 9.674 +time/total (s) 1878.19 +Epoch -821 +---------------------------------- --------------- +2022-05-10 13:42:04.650852 PDT | [0] Epoch -820 finished +---------------------------------- --------------- +epoch -820 +replay_buffer/size 999033 +trainer/num train calls 181000 +trainer/Policy Loss -19.9491 +trainer/Log Pis Mean 24.646 +trainer/Log Pis Std 13.4076 +trainer/Log Pis Max 66.1768 +trainer/Log Pis Min -6.40587 +trainer/policy/mean Mean -0.0325534 +trainer/policy/mean Std 0.902979 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87452 +trainer/policy/normal/std Std 0.590708 +trainer/policy/normal/std Max 5.20917 +trainer/policy/normal/std Min 0.414142 +trainer/policy/normal/log_std Mean 1.02621 +trainer/policy/normal/log_std Std 0.272382 +trainer/policy/normal/log_std Max 1.65042 +trainer/policy/normal/log_std Min -0.881545 +eval/num steps total 180154 +eval/num paths total 181 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.115043 +eval/Actions Std 0.91058 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67011 +time/logging (s) 0.00366064 +time/sampling batch (s) 0.282004 +time/saving (s) 0.00345592 +time/training (s) 7.15423 +time/epoch (s) 10.1135 +time/total (s) 1888.3 +Epoch -820 +---------------------------------- --------------- +2022-05-10 13:42:15.697464 PDT | [0] Epoch -819 finished +---------------------------------- --------------- +epoch -819 +replay_buffer/size 999033 +trainer/num train calls 182000 +trainer/Policy Loss -19.3109 +trainer/Log Pis Mean 23.1965 +trainer/Log Pis Std 12.8797 +trainer/Log Pis Max 60.7143 +trainer/Log Pis Min -11.6214 +trainer/policy/mean Mean -0.0233726 +trainer/policy/mean Std 0.903077 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999972 +trainer/policy/normal/std Mean 2.87428 +trainer/policy/normal/std Std 0.587673 +trainer/policy/normal/std Max 5.16611 +trainer/policy/normal/std Min 0.486614 +trainer/policy/normal/log_std Mean 1.02673 +trainer/policy/normal/log_std Std 0.267707 +trainer/policy/normal/log_std Max 1.64212 +trainer/policy/normal/log_std Min -0.720284 +eval/num steps total 181154 +eval/num paths total 182 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112844 +eval/Actions Std 0.884166 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.9215 +time/logging (s) 0.00371979 +time/sampling batch (s) 0.533258 +time/saving (s) 0.0034749 +time/training (s) 7.56509 +time/epoch (s) 11.027 +time/total (s) 1899.33 +Epoch -819 +---------------------------------- --------------- +2022-05-10 13:42:25.395740 PDT | [0] Epoch -818 finished +---------------------------------- --------------- +epoch -818 +replay_buffer/size 999033 +trainer/num train calls 183000 +trainer/Policy Loss -20.3864 +trainer/Log Pis Mean 24.2497 +trainer/Log Pis Std 12.5961 +trainer/Log Pis Max 63.7146 +trainer/Log Pis Min -7.30321 +trainer/policy/mean Mean -0.0259507 +trainer/policy/mean Std 0.902261 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.86431 +trainer/policy/normal/std Std 0.583382 +trainer/policy/normal/std Max 5.22446 +trainer/policy/normal/std Min 0.455399 +trainer/policy/normal/log_std Mean 1.02423 +trainer/policy/normal/log_std Std 0.261955 +trainer/policy/normal/log_std Max 1.65335 +trainer/policy/normal/log_std Min -0.786581 +eval/num steps total 182154 +eval/num paths total 183 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.44933 +eval/Actions Std 0.87004 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55406 +time/logging (s) 0.00415829 +time/sampling batch (s) 0.2812 +time/saving (s) 0.004425 +time/training (s) 6.83527 +time/epoch (s) 9.67912 +time/total (s) 1909.02 +Epoch -818 +---------------------------------- --------------- +2022-05-10 13:42:35.630220 PDT | [0] Epoch -817 finished +---------------------------------- --------------- +epoch -817 +replay_buffer/size 999033 +trainer/num train calls 184000 +trainer/Policy Loss -18.7809 +trainer/Log Pis Mean 24.4004 +trainer/Log Pis Std 13.0234 +trainer/Log Pis Max 69.0577 +trainer/Log Pis Min -7.29248 +trainer/policy/mean Mean -0.0341243 +trainer/policy/mean Std 0.905224 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7569 +trainer/policy/normal/std Std 0.587636 +trainer/policy/normal/std Max 6.1167 +trainer/policy/normal/std Min 0.425306 +trainer/policy/normal/log_std Mean 0.982273 +trainer/policy/normal/log_std Std 0.280841 +trainer/policy/normal/log_std Max 1.81102 +trainer/policy/normal/log_std Min -0.854946 +eval/num steps total 183154 +eval/num paths total 184 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108722 +eval/Actions Std 0.890093 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72651 +time/logging (s) 0.00386813 +time/sampling batch (s) 0.282979 +time/saving (s) 0.00352307 +time/training (s) 7.19724 +time/epoch (s) 10.2141 +time/total (s) 1919.23 +Epoch -817 +---------------------------------- --------------- +2022-05-10 13:42:45.775090 PDT | [0] Epoch -816 finished +---------------------------------- --------------- +epoch -816 +replay_buffer/size 999033 +trainer/num train calls 185000 +trainer/Policy Loss -19.8715 +trainer/Log Pis Mean 24.8608 +trainer/Log Pis Std 13.4908 +trainer/Log Pis Max 65.8108 +trainer/Log Pis Min -9.86292 +trainer/policy/mean Mean -0.0364167 +trainer/policy/mean Std 0.908308 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.86008 +trainer/policy/normal/std Std 0.563377 +trainer/policy/normal/std Max 5.27499 +trainer/policy/normal/std Min 0.377791 +trainer/policy/normal/log_std Mean 1.02474 +trainer/policy/normal/log_std Std 0.252118 +trainer/policy/normal/log_std Max 1.66298 +trainer/policy/normal/log_std Min -0.973413 +eval/num steps total 184154 +eval/num paths total 185 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0266768 +eval/Actions Std 0.906433 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62775 +time/logging (s) 0.00374595 +time/sampling batch (s) 0.28641 +time/saving (s) 0.0034318 +time/training (s) 7.20372 +time/epoch (s) 10.1251 +time/total (s) 1929.36 +Epoch -816 +---------------------------------- --------------- +2022-05-10 13:42:55.500805 PDT | [0] Epoch -815 finished +---------------------------------- -------------- +epoch -815 +replay_buffer/size 999033 +trainer/num train calls 186000 +trainer/Policy Loss -19.2476 +trainer/Log Pis Mean 24.1172 +trainer/Log Pis Std 12.935 +trainer/Log Pis Max 63.1657 +trainer/Log Pis Min -7.69212 +trainer/policy/mean Mean -0.0484945 +trainer/policy/mean Std 0.90479 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.83777 +trainer/policy/normal/std Std 0.582814 +trainer/policy/normal/std Max 5.27865 +trainer/policy/normal/std Min 0.352639 +trainer/policy/normal/log_std Mean 1.01323 +trainer/policy/normal/log_std Std 0.273696 +trainer/policy/normal/log_std Max 1.66367 +trainer/policy/normal/log_std Min -1.04231 +eval/num steps total 185154 +eval/num paths total 186 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122814 +eval/Actions Std 0.902202 +eval/Actions Max 0.999988 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69631 +time/logging (s) 0.0037664 +time/sampling batch (s) 0.528974 +time/saving (s) 0.0034308 +time/training (s) 6.47369 +time/epoch (s) 9.70617 +time/total (s) 1939.07 +Epoch -815 +---------------------------------- -------------- +2022-05-10 13:43:06.334836 PDT | [0] Epoch -814 finished +---------------------------------- --------------- +epoch -814 +replay_buffer/size 999033 +trainer/num train calls 187000 +trainer/Policy Loss -19.2726 +trainer/Log Pis Mean 25.2793 +trainer/Log Pis Std 12.6625 +trainer/Log Pis Max 64.6315 +trainer/Log Pis Min -6.97278 +trainer/policy/mean Mean -0.029349 +trainer/policy/mean Std 0.906 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.88192 +trainer/policy/normal/std Std 0.589842 +trainer/policy/normal/std Max 6.00488 +trainer/policy/normal/std Min 0.475927 +trainer/policy/normal/log_std Mean 1.02997 +trainer/policy/normal/log_std Std 0.263818 +trainer/policy/normal/log_std Max 1.79257 +trainer/policy/normal/log_std Min -0.742491 +eval/num steps total 186154 +eval/num paths total 187 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0954181 +eval/Actions Std 0.907042 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69074 +time/logging (s) 0.0037155 +time/sampling batch (s) 0.282152 +time/saving (s) 0.00346965 +time/training (s) 7.83431 +time/epoch (s) 10.8144 +time/total (s) 1949.89 +Epoch -814 +---------------------------------- --------------- +2022-05-10 13:43:16.418232 PDT | [0] Epoch -813 finished +---------------------------------- --------------- +epoch -813 +replay_buffer/size 999033 +trainer/num train calls 188000 +trainer/Policy Loss -19.1661 +trainer/Log Pis Mean 24.4444 +trainer/Log Pis Std 12.6058 +trainer/Log Pis Max 63.251 +trainer/Log Pis Min -7.92379 +trainer/policy/mean Mean -0.0485791 +trainer/policy/mean Std 0.900064 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83616 +trainer/policy/normal/std Std 0.603505 +trainer/policy/normal/std Max 6.52001 +trainer/policy/normal/std Min 0.368301 +trainer/policy/normal/log_std Mean 1.01102 +trainer/policy/normal/log_std Std 0.279752 +trainer/policy/normal/log_std Max 1.87488 +trainer/policy/normal/log_std Min -0.998856 +eval/num steps total 187154 +eval/num paths total 188 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0178762 +eval/Actions Std 0.899202 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53354 +time/logging (s) 0.00419442 +time/sampling batch (s) 0.282944 +time/saving (s) 0.00419242 +time/training (s) 7.23929 +time/epoch (s) 10.0642 +time/total (s) 1959.95 +Epoch -813 +---------------------------------- --------------- +2022-05-10 13:43:25.826828 PDT | [0] Epoch -812 finished +---------------------------------- --------------- +epoch -812 +replay_buffer/size 999033 +trainer/num train calls 189000 +trainer/Policy Loss -19.0382 +trainer/Log Pis Mean 23.0627 +trainer/Log Pis Std 12.7608 +trainer/Log Pis Max 69.5702 +trainer/Log Pis Min -9.69247 +trainer/policy/mean Mean -0.021704 +trainer/policy/mean Std 0.900765 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78031 +trainer/policy/normal/std Std 0.570379 +trainer/policy/normal/std Max 4.92218 +trainer/policy/normal/std Min 0.447755 +trainer/policy/normal/log_std Mean 0.99334 +trainer/policy/normal/log_std Std 0.268357 +trainer/policy/normal/log_std Max 1.59375 +trainer/policy/normal/log_std Min -0.80351 +eval/num steps total 188154 +eval/num paths total 189 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0275385 +eval/Actions Std 0.913715 +eval/Actions Max 0.999995 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42348 +time/logging (s) 0.00372061 +time/sampling batch (s) 0.287914 +time/saving (s) 0.00371519 +time/training (s) 6.6688 +time/epoch (s) 9.38763 +time/total (s) 1969.35 +Epoch -812 +---------------------------------- --------------- +2022-05-10 13:43:35.683627 PDT | [0] Epoch -811 finished +---------------------------------- --------------- +epoch -811 +replay_buffer/size 999033 +trainer/num train calls 190000 +trainer/Policy Loss -20.1177 +trainer/Log Pis Mean 25.1457 +trainer/Log Pis Std 13.2854 +trainer/Log Pis Max 63.0264 +trainer/Log Pis Min -8.2013 +trainer/policy/mean Mean -0.0545168 +trainer/policy/mean Std 0.908219 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81931 +trainer/policy/normal/std Std 0.56399 +trainer/policy/normal/std Max 5.03427 +trainer/policy/normal/std Min 0.376803 +trainer/policy/normal/log_std Mean 1.00839 +trainer/policy/normal/log_std Std 0.264921 +trainer/policy/normal/log_std Max 1.61627 +trainer/policy/normal/log_std Min -0.976033 +eval/num steps total 189154 +eval/num paths total 190 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0182807 +eval/Actions Std 0.899171 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46555 +time/logging (s) 0.00377773 +time/sampling batch (s) 0.531693 +time/saving (s) 0.00348124 +time/training (s) 6.83276 +time/epoch (s) 9.83726 +time/total (s) 1979.19 +Epoch -811 +---------------------------------- --------------- +2022-05-10 13:43:45.831274 PDT | [0] Epoch -810 finished +---------------------------------- --------------- +epoch -810 +replay_buffer/size 999033 +trainer/num train calls 191000 +trainer/Policy Loss -19.1677 +trainer/Log Pis Mean 25.0089 +trainer/Log Pis Std 13.8914 +trainer/Log Pis Max 70.9639 +trainer/Log Pis Min -10.7997 +trainer/policy/mean Mean -0.0329472 +trainer/policy/mean Std 0.904274 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86953 +trainer/policy/normal/std Std 0.605648 +trainer/policy/normal/std Max 7.32583 +trainer/policy/normal/std Min 0.376986 +trainer/policy/normal/log_std Mean 1.02305 +trainer/policy/normal/log_std Std 0.278304 +trainer/policy/normal/log_std Max 1.99141 +trainer/policy/normal/log_std Min -0.975547 +eval/num steps total 190154 +eval/num paths total 191 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133483 +eval/Actions Std 0.910787 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71952 +time/logging (s) 0.00382334 +time/sampling batch (s) 0.290185 +time/saving (s) 0.00346526 +time/training (s) 7.11064 +time/epoch (s) 10.1276 +time/total (s) 1989.32 +Epoch -810 +---------------------------------- --------------- +2022-05-10 13:43:55.966868 PDT | [0] Epoch -809 finished +---------------------------------- --------------- +epoch -809 +replay_buffer/size 999033 +trainer/num train calls 192000 +trainer/Policy Loss -19.597 +trainer/Log Pis Mean 23.8414 +trainer/Log Pis Std 13.1748 +trainer/Log Pis Max 71.9093 +trainer/Log Pis Min -6.09993 +trainer/policy/mean Mean -0.0394483 +trainer/policy/mean Std 0.903868 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83072 +trainer/policy/normal/std Std 0.589872 +trainer/policy/normal/std Max 5.06687 +trainer/policy/normal/std Min 0.421216 +trainer/policy/normal/log_std Mean 1.00981 +trainer/policy/normal/log_std Std 0.276643 +trainer/policy/normal/log_std Max 1.62272 +trainer/policy/normal/log_std Min -0.86461 +eval/num steps total 191154 +eval/num paths total 192 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0996687 +eval/Actions Std 0.884545 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61265 +time/logging (s) 0.00394656 +time/sampling batch (s) 0.280807 +time/saving (s) 0.00353724 +time/training (s) 7.21519 +time/epoch (s) 10.1161 +time/total (s) 1999.44 +Epoch -809 +---------------------------------- --------------- +2022-05-10 13:44:05.186243 PDT | [0] Epoch -808 finished +---------------------------------- --------------- +epoch -808 +replay_buffer/size 999033 +trainer/num train calls 193000 +trainer/Policy Loss -19.1623 +trainer/Log Pis Mean 24.5394 +trainer/Log Pis Std 13.2653 +trainer/Log Pis Max 67.4044 +trainer/Log Pis Min -6.22414 +trainer/policy/mean Mean -0.0445732 +trainer/policy/mean Std 0.903545 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77928 +trainer/policy/normal/std Std 0.608305 +trainer/policy/normal/std Max 5.22069 +trainer/policy/normal/std Min 0.393018 +trainer/policy/normal/log_std Mean 0.987314 +trainer/policy/normal/log_std Std 0.297318 +trainer/policy/normal/log_std Max 1.65263 +trainer/policy/normal/log_std Min -0.9339 +eval/num steps total 192154 +eval/num paths total 193 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0553396 +eval/Actions Std 0.90654 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61986 +time/logging (s) 0.00378719 +time/sampling batch (s) 0.279093 +time/saving (s) 0.00360912 +time/training (s) 6.29325 +time/epoch (s) 9.1996 +time/total (s) 2008.64 +Epoch -808 +---------------------------------- --------------- +2022-05-10 13:44:15.604295 PDT | [0] Epoch -807 finished +---------------------------------- --------------- +epoch -807 +replay_buffer/size 999033 +trainer/num train calls 194000 +trainer/Policy Loss -18.4094 +trainer/Log Pis Mean 24.2205 +trainer/Log Pis Std 13.0285 +trainer/Log Pis Max 62.6345 +trainer/Log Pis Min -6.17238 +trainer/policy/mean Mean -0.0237098 +trainer/policy/mean Std 0.901787 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82934 +trainer/policy/normal/std Std 0.590606 +trainer/policy/normal/std Max 5.33448 +trainer/policy/normal/std Min 0.410306 +trainer/policy/normal/log_std Mean 1.00862 +trainer/policy/normal/log_std Std 0.282201 +trainer/policy/normal/log_std Max 1.67419 +trainer/policy/normal/log_std Min -0.890853 +eval/num steps total 193154 +eval/num paths total 194 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.267585 +eval/Actions Std 0.917442 +eval/Actions Max 1 +eval/Actions Min -0.999954 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50485 +time/logging (s) 0.00443948 +time/sampling batch (s) 0.535273 +time/saving (s) 0.00442069 +time/training (s) 7.35001 +time/epoch (s) 10.399 +time/total (s) 2019.04 +Epoch -807 +---------------------------------- --------------- +2022-05-10 13:44:24.518690 PDT | [0] Epoch -806 finished +---------------------------------- --------------- +epoch -806 +replay_buffer/size 999033 +trainer/num train calls 195000 +trainer/Policy Loss -19.2949 +trainer/Log Pis Mean 23.4443 +trainer/Log Pis Std 12.6893 +trainer/Log Pis Max 67.446 +trainer/Log Pis Min -7.61403 +trainer/policy/mean Mean -0.0212615 +trainer/policy/mean Std 0.905448 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86143 +trainer/policy/normal/std Std 0.595344 +trainer/policy/normal/std Max 5.44114 +trainer/policy/normal/std Min 0.342091 +trainer/policy/normal/log_std Mean 1.02121 +trainer/policy/normal/log_std Std 0.273932 +trainer/policy/normal/log_std Max 1.69399 +trainer/policy/normal/log_std Min -1.07268 +eval/num steps total 194154 +eval/num paths total 195 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.236025 +eval/Actions Std 0.892738 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54413 +time/logging (s) 0.00377534 +time/sampling batch (s) 0.278628 +time/saving (s) 0.00344802 +time/training (s) 6.06052 +time/epoch (s) 8.89051 +time/total (s) 2027.94 +Epoch -806 +---------------------------------- --------------- +2022-05-10 13:44:34.807776 PDT | [0] Epoch -805 finished +---------------------------------- --------------- +epoch -805 +replay_buffer/size 999033 +trainer/num train calls 196000 +trainer/Policy Loss -20.5948 +trainer/Log Pis Mean 24.1276 +trainer/Log Pis Std 12.9152 +trainer/Log Pis Max 66.685 +trainer/Log Pis Min -8.39251 +trainer/policy/mean Mean -0.040882 +trainer/policy/mean Std 0.907126 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81259 +trainer/policy/normal/std Std 0.585175 +trainer/policy/normal/std Max 5.22753 +trainer/policy/normal/std Min 0.413137 +trainer/policy/normal/log_std Mean 1.00334 +trainer/policy/normal/log_std Std 0.278053 +trainer/policy/normal/log_std Max 1.65394 +trainer/policy/normal/log_std Min -0.883977 +eval/num steps total 195154 +eval/num paths total 196 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.104631 +eval/Actions Std 0.875397 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61014 +time/logging (s) 0.00369603 +time/sampling batch (s) 0.287676 +time/saving (s) 0.00352444 +time/training (s) 7.36383 +time/epoch (s) 10.2689 +time/total (s) 2038.21 +Epoch -805 +---------------------------------- --------------- +2022-05-10 13:44:45.458365 PDT | [0] Epoch -804 finished +---------------------------------- --------------- +epoch -804 +replay_buffer/size 999033 +trainer/num train calls 197000 +trainer/Policy Loss -19.9952 +trainer/Log Pis Mean 25.6168 +trainer/Log Pis Std 13.7443 +trainer/Log Pis Max 74.2492 +trainer/Log Pis Min -5.59058 +trainer/policy/mean Mean -0.0240168 +trainer/policy/mean Std 0.906318 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83549 +trainer/policy/normal/std Std 0.575719 +trainer/policy/normal/std Max 5.66975 +trainer/policy/normal/std Min 0.391887 +trainer/policy/normal/log_std Mean 1.01395 +trainer/policy/normal/log_std Std 0.26328 +trainer/policy/normal/log_std Max 1.73514 +trainer/policy/normal/log_std Min -0.936781 +eval/num steps total 196154 +eval/num paths total 197 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121962 +eval/Actions Std 0.881497 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6546 +time/logging (s) 0.00377932 +time/sampling batch (s) 0.280691 +time/saving (s) 0.00345593 +time/training (s) 7.68866 +time/epoch (s) 10.6312 +time/total (s) 2048.84 +Epoch -804 +---------------------------------- --------------- +2022-05-10 13:44:54.923332 PDT | [0] Epoch -803 finished +---------------------------------- --------------- +epoch -803 +replay_buffer/size 999033 +trainer/num train calls 198000 +trainer/Policy Loss -19.3361 +trainer/Log Pis Mean 24.1559 +trainer/Log Pis Std 12.6949 +trainer/Log Pis Max 64.9846 +trainer/Log Pis Min -6.72637 +trainer/policy/mean Mean -0.0462145 +trainer/policy/mean Std 0.905314 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.77533 +trainer/policy/normal/std Std 0.578836 +trainer/policy/normal/std Max 5.82211 +trainer/policy/normal/std Min 0.407662 +trainer/policy/normal/log_std Mean 0.991388 +trainer/policy/normal/log_std Std 0.267608 +trainer/policy/normal/log_std Max 1.76166 +trainer/policy/normal/log_std Min -0.897318 +eval/num steps total 197154 +eval/num paths total 198 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0739969 +eval/Actions Std 0.919993 +eval/Actions Max 0.999986 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66624 +time/logging (s) 0.00378995 +time/sampling batch (s) 0.278155 +time/saving (s) 0.00366448 +time/training (s) 6.4938 +time/epoch (s) 9.44566 +time/total (s) 2058.29 +Epoch -803 +---------------------------------- --------------- +2022-05-10 13:45:04.403129 PDT | [0] Epoch -802 finished +---------------------------------- --------------- +epoch -802 +replay_buffer/size 999033 +trainer/num train calls 199000 +trainer/Policy Loss -19.9994 +trainer/Log Pis Mean 24.6708 +trainer/Log Pis Std 12.9317 +trainer/Log Pis Max 68.686 +trainer/Log Pis Min -7.579 +trainer/policy/mean Mean -0.0330621 +trainer/policy/mean Std 0.907441 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7946 +trainer/policy/normal/std Std 0.596029 +trainer/policy/normal/std Max 5.6737 +trainer/policy/normal/std Min 0.45854 +trainer/policy/normal/log_std Mean 0.997121 +trainer/policy/normal/log_std Std 0.272057 +trainer/policy/normal/log_std Max 1.73584 +trainer/policy/normal/log_std Min -0.779708 +eval/num steps total 198154 +eval/num paths total 199 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.384919 +eval/Actions Std 0.826225 +eval/Actions Max 0.999985 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83642 +time/logging (s) 0.00409584 +time/sampling batch (s) 0.279961 +time/saving (s) 0.00413298 +time/training (s) 6.33586 +time/epoch (s) 9.46047 +time/total (s) 2067.75 +Epoch -802 +---------------------------------- --------------- +2022-05-10 13:45:14.308922 PDT | [0] Epoch -801 finished +---------------------------------- --------------- +epoch -801 +replay_buffer/size 999033 +trainer/num train calls 200000 +trainer/Policy Loss -21.0129 +trainer/Log Pis Mean 24.7707 +trainer/Log Pis Std 13.5293 +trainer/Log Pis Max 77.6828 +trainer/Log Pis Min -9.35289 +trainer/policy/mean Mean -0.0247125 +trainer/policy/mean Std 0.909169 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.88828 +trainer/policy/normal/std Std 0.582206 +trainer/policy/normal/std Max 5.66489 +trainer/policy/normal/std Min 0.466056 +trainer/policy/normal/log_std Mean 1.03372 +trainer/policy/normal/log_std Std 0.254622 +trainer/policy/normal/log_std Max 1.73429 +trainer/policy/normal/log_std Min -0.763449 +eval/num steps total 199154 +eval/num paths total 200 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0551583 +eval/Actions Std 0.916116 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73745 +time/logging (s) 0.00370419 +time/sampling batch (s) 0.287998 +time/saving (s) 0.00664388 +time/training (s) 6.84936 +time/epoch (s) 9.88515 +time/total (s) 2077.64 +Epoch -801 +---------------------------------- --------------- +2022-05-10 13:45:24.661013 PDT | [0] Epoch -800 finished +---------------------------------- --------------- +epoch -800 +replay_buffer/size 999033 +trainer/num train calls 201000 +trainer/Policy Loss -19.483 +trainer/Log Pis Mean 24.0598 +trainer/Log Pis Std 12.5285 +trainer/Log Pis Max 65.7407 +trainer/Log Pis Min -5.01268 +trainer/policy/mean Mean -0.0491534 +trainer/policy/mean Std 0.902799 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85194 +trainer/policy/normal/std Std 0.588274 +trainer/policy/normal/std Max 5.36086 +trainer/policy/normal/std Min 0.407484 +trainer/policy/normal/log_std Mean 1.01848 +trainer/policy/normal/log_std Std 0.270587 +trainer/policy/normal/log_std Max 1.67912 +trainer/policy/normal/log_std Min -0.897755 +eval/num steps total 200154 +eval/num paths total 201 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179485 +eval/Actions Std 0.852493 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85362 +time/logging (s) 0.00368947 +time/sampling batch (s) 0.281876 +time/saving (s) 0.00345915 +time/training (s) 7.18967 +time/epoch (s) 10.3323 +time/total (s) 2087.98 +Epoch -800 +---------------------------------- --------------- +2022-05-10 13:45:35.511893 PDT | [0] Epoch -799 finished +---------------------------------- --------------- +epoch -799 +replay_buffer/size 999033 +trainer/num train calls 202000 +trainer/Policy Loss -19.6932 +trainer/Log Pis Mean 24.9476 +trainer/Log Pis Std 13.1839 +trainer/Log Pis Max 63.2428 +trainer/Log Pis Min -6.78147 +trainer/policy/mean Mean -0.0275301 +trainer/policy/mean Std 0.906893 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85425 +trainer/policy/normal/std Std 0.569483 +trainer/policy/normal/std Max 4.86557 +trainer/policy/normal/std Min 0.380415 +trainer/policy/normal/log_std Mean 1.02126 +trainer/policy/normal/log_std Std 0.260864 +trainer/policy/normal/log_std Max 1.58218 +trainer/policy/normal/log_std Min -0.966492 +eval/num steps total 201154 +eval/num paths total 202 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.212428 +eval/Actions Std 0.957304 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65559 +time/logging (s) 0.00373142 +time/sampling batch (s) 0.538078 +time/saving (s) 0.00350078 +time/training (s) 7.63001 +time/epoch (s) 10.8309 +time/total (s) 2098.81 +Epoch -799 +---------------------------------- --------------- +2022-05-10 13:45:47.041511 PDT | [0] Epoch -798 finished +---------------------------------- --------------- +epoch -798 +replay_buffer/size 999033 +trainer/num train calls 203000 +trainer/Policy Loss -19.3525 +trainer/Log Pis Mean 24.1913 +trainer/Log Pis Std 12.2189 +trainer/Log Pis Max 64.6087 +trainer/Log Pis Min -5.04225 +trainer/policy/mean Mean -0.0181362 +trainer/policy/mean Std 0.90594 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.82265 +trainer/policy/normal/std Std 0.589561 +trainer/policy/normal/std Max 5.95103 +trainer/policy/normal/std Min 0.436756 +trainer/policy/normal/log_std Mean 1.00756 +trainer/policy/normal/log_std Std 0.271932 +trainer/policy/normal/log_std Max 1.78356 +trainer/policy/normal/log_std Min -0.828382 +eval/num steps total 202154 +eval/num paths total 203 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.233116 +eval/Actions Std 0.913489 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64178 +time/logging (s) 0.00379496 +time/sampling batch (s) 1.0337 +time/saving (s) 0.00349539 +time/training (s) 7.8271 +time/epoch (s) 11.5099 +time/total (s) 2110.32 +Epoch -798 +---------------------------------- --------------- +2022-05-10 13:45:56.841803 PDT | [0] Epoch -797 finished +---------------------------------- --------------- +epoch -797 +replay_buffer/size 999033 +trainer/num train calls 204000 +trainer/Policy Loss -18.5606 +trainer/Log Pis Mean 25.1561 +trainer/Log Pis Std 13.935 +trainer/Log Pis Max 64.5195 +trainer/Log Pis Min -8.83854 +trainer/policy/mean Mean -0.0334802 +trainer/policy/mean Std 0.90679 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.88699 +trainer/policy/normal/std Std 0.590796 +trainer/policy/normal/std Max 5.79484 +trainer/policy/normal/std Min 0.466247 +trainer/policy/normal/log_std Mean 1.03161 +trainer/policy/normal/log_std Std 0.264493 +trainer/policy/normal/log_std Max 1.75697 +trainer/policy/normal/log_std Min -0.76304 +eval/num steps total 203154 +eval/num paths total 204 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.426564 +eval/Actions Std 0.855841 +eval/Actions Max 0.999985 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62623 +time/logging (s) 0.00377062 +time/sampling batch (s) 0.284698 +time/saving (s) 0.0035014 +time/training (s) 6.86221 +time/epoch (s) 9.78042 +time/total (s) 2120.11 +Epoch -797 +---------------------------------- --------------- +2022-05-10 13:46:06.980033 PDT | [0] Epoch -796 finished +---------------------------------- --------------- +epoch -796 +replay_buffer/size 999033 +trainer/num train calls 205000 +trainer/Policy Loss -20.0046 +trainer/Log Pis Mean 25.949 +trainer/Log Pis Std 13.3225 +trainer/Log Pis Max 77.0311 +trainer/Log Pis Min -6.03893 +trainer/policy/mean Mean -0.0338928 +trainer/policy/mean Std 0.90784 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.90428 +trainer/policy/normal/std Std 0.620985 +trainer/policy/normal/std Max 6.10698 +trainer/policy/normal/std Min 0.3973 +trainer/policy/normal/log_std Mean 1.03414 +trainer/policy/normal/log_std Std 0.28382 +trainer/policy/normal/log_std Max 1.80943 +trainer/policy/normal/log_std Min -0.923064 +eval/num steps total 204154 +eval/num paths total 205 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.140625 +eval/Actions Std 0.890892 +eval/Actions Max 0.99999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64047 +time/logging (s) 0.00424938 +time/sampling batch (s) 0.281479 +time/saving (s) 0.00449886 +time/training (s) 7.18818 +time/epoch (s) 10.1189 +time/total (s) 2130.23 +Epoch -796 +---------------------------------- --------------- +2022-05-10 13:46:17.002139 PDT | [0] Epoch -795 finished +---------------------------------- --------------- +epoch -795 +replay_buffer/size 999033 +trainer/num train calls 206000 +trainer/Policy Loss -18.7022 +trainer/Log Pis Mean 24.951 +trainer/Log Pis Std 12.8416 +trainer/Log Pis Max 64.2996 +trainer/Log Pis Min -9.93548 +trainer/policy/mean Mean -0.0295165 +trainer/policy/mean Std 0.906901 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87347 +trainer/policy/normal/std Std 0.597329 +trainer/policy/normal/std Max 5.15509 +trainer/policy/normal/std Min 0.291505 +trainer/policy/normal/log_std Mean 1.02452 +trainer/policy/normal/log_std Std 0.28061 +trainer/policy/normal/log_std Max 1.63998 +trainer/policy/normal/log_std Min -1.2327 +eval/num steps total 205154 +eval/num paths total 206 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138684 +eval/Actions Std 0.882834 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56185 +time/logging (s) 0.00371432 +time/sampling batch (s) 0.285163 +time/saving (s) 0.00349701 +time/training (s) 7.14696 +time/epoch (s) 10.0012 +time/total (s) 2140.23 +Epoch -795 +---------------------------------- --------------- +2022-05-10 13:46:27.849274 PDT | [0] Epoch -794 finished +---------------------------------- --------------- +epoch -794 +replay_buffer/size 999033 +trainer/num train calls 207000 +trainer/Policy Loss -21.156 +trainer/Log Pis Mean 24.9936 +trainer/Log Pis Std 12.103 +trainer/Log Pis Max 67.657 +trainer/Log Pis Min -8.01057 +trainer/policy/mean Mean -0.0207608 +trainer/policy/mean Std 0.904003 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80179 +trainer/policy/normal/std Std 0.586612 +trainer/policy/normal/std Max 6.93441 +trainer/policy/normal/std Min 0.402495 +trainer/policy/normal/log_std Mean 0.999893 +trainer/policy/normal/log_std Std 0.274669 +trainer/policy/normal/log_std Max 1.9365 +trainer/policy/normal/log_std Min -0.910072 +eval/num steps total 206154 +eval/num paths total 207 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.15326 +eval/Actions Std 0.872815 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79082 +time/logging (s) 0.00370986 +time/sampling batch (s) 0.282415 +time/saving (s) 0.00347629 +time/training (s) 7.74693 +time/epoch (s) 10.8274 +time/total (s) 2151.06 +Epoch -794 +---------------------------------- --------------- +2022-05-10 13:46:38.786076 PDT | [0] Epoch -793 finished +---------------------------------- --------------- +epoch -793 +replay_buffer/size 999033 +trainer/num train calls 208000 +trainer/Policy Loss -19.7205 +trainer/Log Pis Mean 23.7843 +trainer/Log Pis Std 13.3062 +trainer/Log Pis Max 74.2079 +trainer/Log Pis Min -9.39641 +trainer/policy/mean Mean -0.0249312 +trainer/policy/mean Std 0.900945 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81058 +trainer/policy/normal/std Std 0.574093 +trainer/policy/normal/std Max 4.74661 +trainer/policy/normal/std Min 0.505382 +trainer/policy/normal/log_std Mean 1.00505 +trainer/policy/normal/log_std Std 0.262957 +trainer/policy/normal/log_std Max 1.55743 +trainer/policy/normal/log_std Min -0.682441 +eval/num steps total 207154 +eval/num paths total 208 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0179841 +eval/Actions Std 0.901556 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61895 +time/logging (s) 0.00381821 +time/sampling batch (s) 0.282454 +time/saving (s) 0.00341997 +time/training (s) 8.00856 +time/epoch (s) 10.9172 +time/total (s) 2161.98 +Epoch -793 +---------------------------------- --------------- +2022-05-10 13:46:49.137013 PDT | [0] Epoch -792 finished +---------------------------------- --------------- +epoch -792 +replay_buffer/size 999033 +trainer/num train calls 209000 +trainer/Policy Loss -18.6302 +trainer/Log Pis Mean 24.1335 +trainer/Log Pis Std 13.8961 +trainer/Log Pis Max 64.684 +trainer/Log Pis Min -7.86508 +trainer/policy/mean Mean -0.0314972 +trainer/policy/mean Std 0.908563 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.84723 +trainer/policy/normal/std Std 0.586703 +trainer/policy/normal/std Max 6.04735 +trainer/policy/normal/std Min 0.445184 +trainer/policy/normal/log_std Mean 1.01641 +trainer/policy/normal/log_std Std 0.273414 +trainer/policy/normal/log_std Max 1.79962 +trainer/policy/normal/log_std Min -0.809267 +eval/num steps total 208154 +eval/num paths total 209 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130908 +eval/Actions Std 0.878799 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59408 +time/logging (s) 0.00368496 +time/sampling batch (s) 0.282408 +time/saving (s) 0.0034461 +time/training (s) 7.44731 +time/epoch (s) 10.3309 +time/total (s) 2172.32 +Epoch -792 +---------------------------------- --------------- +2022-05-10 13:46:59.774571 PDT | [0] Epoch -791 finished +---------------------------------- --------------- +epoch -791 +replay_buffer/size 999033 +trainer/num train calls 210000 +trainer/Policy Loss -19.584 +trainer/Log Pis Mean 24.5363 +trainer/Log Pis Std 12.9798 +trainer/Log Pis Max 61.703 +trainer/Log Pis Min -10.9125 +trainer/policy/mean Mean -0.0466501 +trainer/policy/mean Std 0.901262 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.91144 +trainer/policy/normal/std Std 0.593503 +trainer/policy/normal/std Max 5.40071 +trainer/policy/normal/std Min 0.445047 +trainer/policy/normal/log_std Mean 1.04053 +trainer/policy/normal/log_std Std 0.261812 +trainer/policy/normal/log_std Max 1.68653 +trainer/policy/normal/log_std Min -0.809574 +eval/num steps total 209154 +eval/num paths total 210 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.139337 +eval/Actions Std 0.895165 +eval/Actions Max 0.99999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62477 +time/logging (s) 0.00404803 +time/sampling batch (s) 0.53325 +time/saving (s) 0.00407449 +time/training (s) 7.45196 +time/epoch (s) 10.6181 +time/total (s) 2182.94 +Epoch -791 +---------------------------------- --------------- +2022-05-10 13:47:09.833317 PDT | [0] Epoch -790 finished +---------------------------------- --------------- +epoch -790 +replay_buffer/size 999033 +trainer/num train calls 211000 +trainer/Policy Loss -19.6188 +trainer/Log Pis Mean 25.4676 +trainer/Log Pis Std 13.6303 +trainer/Log Pis Max 79.9789 +trainer/Log Pis Min -8.67518 +trainer/policy/mean Mean -0.024962 +trainer/policy/mean Std 0.905802 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84807 +trainer/policy/normal/std Std 0.59179 +trainer/policy/normal/std Max 5.62833 +trainer/policy/normal/std Min 0.34503 +trainer/policy/normal/log_std Mean 1.01679 +trainer/policy/normal/log_std Std 0.27117 +trainer/policy/normal/log_std Max 1.72781 +trainer/policy/normal/log_std Min -1.06412 +eval/num steps total 210154 +eval/num paths total 211 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.372336 +eval/Actions Std 0.893817 +eval/Actions Max 0.999998 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82336 +time/logging (s) 0.00382802 +time/sampling batch (s) 0.282322 +time/saving (s) 0.00363243 +time/training (s) 6.92532 +time/epoch (s) 10.0385 +time/total (s) 2192.98 +Epoch -790 +---------------------------------- --------------- +2022-05-10 13:47:19.724115 PDT | [0] Epoch -789 finished +---------------------------------- --------------- +epoch -789 +replay_buffer/size 999033 +trainer/num train calls 212000 +trainer/Policy Loss -17.5759 +trainer/Log Pis Mean 23.9371 +trainer/Log Pis Std 12.8902 +trainer/Log Pis Max 60.8911 +trainer/Log Pis Min -11.3329 +trainer/policy/mean Mean -0.00596348 +trainer/policy/mean Std 0.899125 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7967 +trainer/policy/normal/std Std 0.571014 +trainer/policy/normal/std Max 5.1651 +trainer/policy/normal/std Min 0.398056 +trainer/policy/normal/log_std Mean 0.999565 +trainer/policy/normal/log_std Std 0.266899 +trainer/policy/normal/log_std Max 1.64192 +trainer/policy/normal/log_std Min -0.921163 +eval/num steps total 211154 +eval/num paths total 212 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00802093 +eval/Actions Std 0.981181 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40634 +time/logging (s) 0.00376971 +time/sampling batch (s) 0.285572 +time/saving (s) 0.00346282 +time/training (s) 7.17145 +time/epoch (s) 9.8706 +time/total (s) 2202.85 +Epoch -789 +---------------------------------- --------------- +2022-05-10 13:47:29.674083 PDT | [0] Epoch -788 finished +---------------------------------- --------------- +epoch -788 +replay_buffer/size 999033 +trainer/num train calls 213000 +trainer/Policy Loss -20.0052 +trainer/Log Pis Mean 25.3989 +trainer/Log Pis Std 13.7482 +trainer/Log Pis Max 72.0795 +trainer/Log Pis Min -12.2622 +trainer/policy/mean Mean -0.0392745 +trainer/policy/mean Std 0.906528 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.89854 +trainer/policy/normal/std Std 0.584606 +trainer/policy/normal/std Max 5.23579 +trainer/policy/normal/std Min 0.438609 +trainer/policy/normal/log_std Mean 1.03625 +trainer/policy/normal/log_std Std 0.261989 +trainer/policy/normal/log_std Max 1.65552 +trainer/policy/normal/log_std Min -0.824148 +eval/num steps total 212154 +eval/num paths total 213 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0586516 +eval/Actions Std 0.907994 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5733 +time/logging (s) 0.00369008 +time/sampling batch (s) 0.282387 +time/saving (s) 0.00345229 +time/training (s) 7.06742 +time/epoch (s) 9.93025 +time/total (s) 2212.78 +Epoch -788 +---------------------------------- --------------- +2022-05-10 13:47:40.424196 PDT | [0] Epoch -787 finished +---------------------------------- ---------------- +epoch -787 +replay_buffer/size 999033 +trainer/num train calls 214000 +trainer/Policy Loss -18.4859 +trainer/Log Pis Mean 24.0745 +trainer/Log Pis Std 12.6062 +trainer/Log Pis Max 60.5483 +trainer/Log Pis Min -3.00755 +trainer/policy/mean Mean -0.0446959 +trainer/policy/mean Std 0.904207 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79361 +trainer/policy/normal/std Std 0.587732 +trainer/policy/normal/std Max 5.27187 +trainer/policy/normal/std Min 0.427927 +trainer/policy/normal/log_std Mean 0.99667 +trainer/policy/normal/log_std Std 0.275227 +trainer/policy/normal/log_std Max 1.66238 +trainer/policy/normal/log_std Min -0.848802 +eval/num steps total 213154 +eval/num paths total 214 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000203965 +eval/Actions Std 0.958967 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46256 +time/logging (s) 0.00369994 +time/sampling batch (s) 0.532728 +time/saving (s) 0.00344483 +time/training (s) 7.72788 +time/epoch (s) 10.7303 +time/total (s) 2223.52 +Epoch -787 +---------------------------------- ---------------- +2022-05-10 13:47:50.659481 PDT | [0] Epoch -786 finished +---------------------------------- --------------- +epoch -786 +replay_buffer/size 999033 +trainer/num train calls 215000 +trainer/Policy Loss -19.2927 +trainer/Log Pis Mean 24.1402 +trainer/Log Pis Std 13.2028 +trainer/Log Pis Max 76.9307 +trainer/Log Pis Min -4.29299 +trainer/policy/mean Mean -0.0251206 +trainer/policy/mean Std 0.908255 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85287 +trainer/policy/normal/std Std 0.579388 +trainer/policy/normal/std Max 5.04504 +trainer/policy/normal/std Min 0.400473 +trainer/policy/normal/log_std Mean 1.02002 +trainer/policy/normal/log_std Std 0.263575 +trainer/policy/normal/log_std Max 1.61841 +trainer/policy/normal/log_std Min -0.915108 +eval/num steps total 214154 +eval/num paths total 215 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0421659 +eval/Actions Std 0.894463 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63214 +time/logging (s) 0.00422064 +time/sampling batch (s) 0.527635 +time/saving (s) 0.00409659 +time/training (s) 7.04829 +time/epoch (s) 10.2164 +time/total (s) 2233.74 +Epoch -786 +---------------------------------- --------------- +2022-05-10 13:48:01.003328 PDT | [0] Epoch -785 finished +---------------------------------- --------------- +epoch -785 +replay_buffer/size 999033 +trainer/num train calls 216000 +trainer/Policy Loss -20.154 +trainer/Log Pis Mean 24.6704 +trainer/Log Pis Std 13.2511 +trainer/Log Pis Max 71.1378 +trainer/Log Pis Min -10.4183 +trainer/policy/mean Mean -0.0258074 +trainer/policy/mean Std 0.905883 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.89374 +trainer/policy/normal/std Std 0.580817 +trainer/policy/normal/std Max 6.32631 +trainer/policy/normal/std Min 0.40111 +trainer/policy/normal/log_std Mean 1.03427 +trainer/policy/normal/log_std Std 0.265821 +trainer/policy/normal/log_std Max 1.84472 +trainer/policy/normal/log_std Min -0.913519 +eval/num steps total 215154 +eval/num paths total 216 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.1005 +eval/Actions Std 0.873009 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58168 +time/logging (s) 0.00376278 +time/sampling batch (s) 0.528155 +time/saving (s) 0.0035597 +time/training (s) 7.20638 +time/epoch (s) 10.3235 +time/total (s) 2244.06 +Epoch -785 +---------------------------------- --------------- +2022-05-10 13:48:10.988094 PDT | [0] Epoch -784 finished +---------------------------------- --------------- +epoch -784 +replay_buffer/size 999033 +trainer/num train calls 217000 +trainer/Policy Loss -18.8758 +trainer/Log Pis Mean 25.3 +trainer/Log Pis Std 13.4842 +trainer/Log Pis Max 71.2817 +trainer/Log Pis Min -6.72168 +trainer/policy/mean Mean -0.0285855 +trainer/policy/mean Std 0.908193 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89708 +trainer/policy/normal/std Std 0.593541 +trainer/policy/normal/std Max 5.92785 +trainer/policy/normal/std Min 0.429178 +trainer/policy/normal/log_std Mean 1.03529 +trainer/policy/normal/log_std Std 0.263479 +trainer/policy/normal/log_std Max 1.77966 +trainer/policy/normal/log_std Min -0.845883 +eval/num steps total 216154 +eval/num paths total 217 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0546664 +eval/Actions Std 0.878197 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58473 +time/logging (s) 0.00369778 +time/sampling batch (s) 0.277395 +time/saving (s) 0.00356188 +time/training (s) 7.09589 +time/epoch (s) 9.96527 +time/total (s) 2254.03 +Epoch -784 +---------------------------------- --------------- +2022-05-10 13:48:22.694780 PDT | [0] Epoch -783 finished +---------------------------------- --------------- +epoch -783 +replay_buffer/size 999033 +trainer/num train calls 218000 +trainer/Policy Loss -19.1964 +trainer/Log Pis Mean 24.1195 +trainer/Log Pis Std 13.2251 +trainer/Log Pis Max 71.451 +trainer/Log Pis Min -13.6699 +trainer/policy/mean Mean -0.0333978 +trainer/policy/mean Std 0.905336 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.88698 +trainer/policy/normal/std Std 0.589081 +trainer/policy/normal/std Max 5.82477 +trainer/policy/normal/std Min 0.421289 +trainer/policy/normal/log_std Mean 1.03173 +trainer/policy/normal/log_std Std 0.265498 +trainer/policy/normal/log_std Max 1.76212 +trainer/policy/normal/log_std Min -0.864436 +eval/num steps total 217154 +eval/num paths total 218 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0811161 +eval/Actions Std 0.897458 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61099 +time/logging (s) 0.00369221 +time/sampling batch (s) 1.52965 +time/saving (s) 0.00340641 +time/training (s) 7.53954 +time/epoch (s) 11.6873 +time/total (s) 2265.72 +Epoch -783 +---------------------------------- --------------- +2022-05-10 13:48:33.225936 PDT | [0] Epoch -782 finished +---------------------------------- --------------- +epoch -782 +replay_buffer/size 999033 +trainer/num train calls 219000 +trainer/Policy Loss -19.3634 +trainer/Log Pis Mean 23.6731 +trainer/Log Pis Std 13.4374 +trainer/Log Pis Max 75.5187 +trainer/Log Pis Min -4.88435 +trainer/policy/mean Mean -0.0429389 +trainer/policy/mean Std 0.901937 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84383 +trainer/policy/normal/std Std 0.598202 +trainer/policy/normal/std Max 4.99163 +trainer/policy/normal/std Min 0.355767 +trainer/policy/normal/log_std Mean 1.0134 +trainer/policy/normal/log_std Std 0.283466 +trainer/policy/normal/log_std Max 1.60776 +trainer/policy/normal/log_std Min -1.03348 +eval/num steps total 218154 +eval/num paths total 219 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.293509 +eval/Actions Std 0.891803 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73493 +time/logging (s) 0.00373248 +time/sampling batch (s) 0.279949 +time/saving (s) 0.00342669 +time/training (s) 7.48958 +time/epoch (s) 10.5116 +time/total (s) 2276.24 +Epoch -782 +---------------------------------- --------------- +2022-05-10 13:48:44.301813 PDT | [0] Epoch -781 finished +---------------------------------- --------------- +epoch -781 +replay_buffer/size 999033 +trainer/num train calls 220000 +trainer/Policy Loss -20.6206 +trainer/Log Pis Mean 24.0798 +trainer/Log Pis Std 13.1547 +trainer/Log Pis Max 70.6441 +trainer/Log Pis Min -8.76145 +trainer/policy/mean Mean -0.0447936 +trainer/policy/mean Std 0.909683 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85102 +trainer/policy/normal/std Std 0.576137 +trainer/policy/normal/std Max 5.60431 +trainer/policy/normal/std Min 0.38646 +trainer/policy/normal/log_std Mean 1.02019 +trainer/policy/normal/log_std Std 0.258293 +trainer/policy/normal/log_std Max 1.72354 +trainer/policy/normal/log_std Min -0.950727 +eval/num steps total 219154 +eval/num paths total 220 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.183083 +eval/Actions Std 0.954005 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6795 +time/logging (s) 0.00419154 +time/sampling batch (s) 0.536772 +time/saving (s) 0.00412241 +time/training (s) 7.83163 +time/epoch (s) 11.0562 +time/total (s) 2287.3 +Epoch -781 +---------------------------------- --------------- +2022-05-10 13:48:54.087094 PDT | [0] Epoch -780 finished +---------------------------------- --------------- +epoch -780 +replay_buffer/size 999033 +trainer/num train calls 221000 +trainer/Policy Loss -20.5801 +trainer/Log Pis Mean 24.6424 +trainer/Log Pis Std 13.315 +trainer/Log Pis Max 66.7763 +trainer/Log Pis Min -8.56742 +trainer/policy/mean Mean -0.047672 +trainer/policy/mean Std 0.906837 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.875 +trainer/policy/normal/std Std 0.615538 +trainer/policy/normal/std Max 5.35188 +trainer/policy/normal/std Min 0.397899 +trainer/policy/normal/log_std Mean 1.02363 +trainer/policy/normal/log_std Std 0.285344 +trainer/policy/normal/log_std Max 1.67745 +trainer/policy/normal/log_std Min -0.921558 +eval/num steps total 220154 +eval/num paths total 221 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115609 +eval/Actions Std 0.90478 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45734 +time/logging (s) 0.00377892 +time/sampling batch (s) 0.531972 +time/saving (s) 0.00381423 +time/training (s) 6.76749 +time/epoch (s) 9.76439 +time/total (s) 2297.06 +Epoch -780 +---------------------------------- --------------- +2022-05-10 13:49:04.639304 PDT | [0] Epoch -779 finished +---------------------------------- --------------- +epoch -779 +replay_buffer/size 999033 +trainer/num train calls 222000 +trainer/Policy Loss -19.2152 +trainer/Log Pis Mean 24.8855 +trainer/Log Pis Std 13.486 +trainer/Log Pis Max 74.795 +trainer/Log Pis Min -6.39316 +trainer/policy/mean Mean -0.0283332 +trainer/policy/mean Std 0.904352 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84172 +trainer/policy/normal/std Std 0.588855 +trainer/policy/normal/std Max 5.29759 +trainer/policy/normal/std Min 0.478974 +trainer/policy/normal/log_std Mean 1.01561 +trainer/policy/normal/log_std Std 0.264115 +trainer/policy/normal/log_std Max 1.66725 +trainer/policy/normal/log_std Min -0.736109 +eval/num steps total 221154 +eval/num paths total 222 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00939038 +eval/Actions Std 0.898474 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70169 +time/logging (s) 0.00373314 +time/sampling batch (s) 0.283256 +time/saving (s) 0.0034909 +time/training (s) 7.53977 +time/epoch (s) 10.5319 +time/total (s) 2307.6 +Epoch -779 +---------------------------------- --------------- +2022-05-10 13:49:16.259882 PDT | [0] Epoch -778 finished +---------------------------------- --------------- +epoch -778 +replay_buffer/size 999033 +trainer/num train calls 223000 +trainer/Policy Loss -19.8981 +trainer/Log Pis Mean 25.6387 +trainer/Log Pis Std 13.1476 +trainer/Log Pis Max 73.7278 +trainer/Log Pis Min -9.75386 +trainer/policy/mean Mean -0.0470194 +trainer/policy/mean Std 0.903962 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86943 +trainer/policy/normal/std Std 0.627622 +trainer/policy/normal/std Max 5.96045 +trainer/policy/normal/std Min 0.348486 +trainer/policy/normal/log_std Mean 1.01988 +trainer/policy/normal/log_std Std 0.294341 +trainer/policy/normal/log_std Max 1.78515 +trainer/policy/normal/log_std Min -1.05416 +eval/num steps total 222154 +eval/num paths total 223 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0132108 +eval/Actions Std 0.971178 +eval/Actions Max 0.999988 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71419 +time/logging (s) 0.00369666 +time/sampling batch (s) 0.783343 +time/saving (s) 0.00356741 +time/training (s) 8.09596 +time/epoch (s) 11.6008 +time/total (s) 2319.2 +Epoch -778 +---------------------------------- --------------- +2022-05-10 13:49:26.616388 PDT | [0] Epoch -777 finished +---------------------------------- --------------- +epoch -777 +replay_buffer/size 999033 +trainer/num train calls 224000 +trainer/Policy Loss -20.3566 +trainer/Log Pis Mean 24.4379 +trainer/Log Pis Std 13.698 +trainer/Log Pis Max 73.4067 +trainer/Log Pis Min -11.4753 +trainer/policy/mean Mean -0.0557899 +trainer/policy/mean Std 0.90956 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83647 +trainer/policy/normal/std Std 0.586954 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.389814 +trainer/policy/normal/log_std Mean 1.01326 +trainer/policy/normal/log_std Std 0.269826 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -0.942086 +eval/num steps total 223154 +eval/num paths total 224 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0170228 +eval/Actions Std 0.974998 +eval/Actions Max 0.999978 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42442 +time/logging (s) 0.00373937 +time/sampling batch (s) 0.283544 +time/saving (s) 0.00346942 +time/training (s) 7.62141 +time/epoch (s) 10.3366 +time/total (s) 2329.54 +Epoch -777 +---------------------------------- --------------- +2022-05-10 13:49:36.128472 PDT | [0] Epoch -776 finished +---------------------------------- --------------- +epoch -776 +replay_buffer/size 999033 +trainer/num train calls 225000 +trainer/Policy Loss -20.2806 +trainer/Log Pis Mean 25.273 +trainer/Log Pis Std 13.5319 +trainer/Log Pis Max 76.7145 +trainer/Log Pis Min -7.41753 +trainer/policy/mean Mean -0.0340353 +trainer/policy/mean Std 0.907419 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85983 +trainer/policy/normal/std Std 0.616801 +trainer/policy/normal/std Max 5.23692 +trainer/policy/normal/std Min 0.368152 +trainer/policy/normal/log_std Mean 1.01834 +trainer/policy/normal/log_std Std 0.28437 +trainer/policy/normal/log_std Max 1.65573 +trainer/policy/normal/log_std Min -0.99926 +eval/num steps total 224154 +eval/num paths total 225 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0566354 +eval/Actions Std 0.91286 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65909 +time/logging (s) 0.00374621 +time/sampling batch (s) 0.283057 +time/saving (s) 0.00350475 +time/training (s) 6.54276 +time/epoch (s) 9.49216 +time/total (s) 2339.04 +Epoch -776 +---------------------------------- --------------- +2022-05-10 13:49:47.009146 PDT | [0] Epoch -775 finished +---------------------------------- --------------- +epoch -775 +replay_buffer/size 999033 +trainer/num train calls 226000 +trainer/Policy Loss -19.3039 +trainer/Log Pis Mean 24.0105 +trainer/Log Pis Std 12.9782 +trainer/Log Pis Max 65.6362 +trainer/Log Pis Min -6.93036 +trainer/policy/mean Mean -0.0467106 +trainer/policy/mean Std 0.903645 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.79983 +trainer/policy/normal/std Std 0.591392 +trainer/policy/normal/std Max 5.78088 +trainer/policy/normal/std Min 0.401742 +trainer/policy/normal/log_std Mean 0.997664 +trainer/policy/normal/log_std Std 0.283116 +trainer/policy/normal/log_std Max 1.75456 +trainer/policy/normal/log_std Min -0.911945 +eval/num steps total 225154 +eval/num paths total 226 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0359558 +eval/Actions Std 0.95996 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77062 +time/logging (s) 0.00410949 +time/sampling batch (s) 0.784052 +time/saving (s) 0.00421452 +time/training (s) 7.29807 +time/epoch (s) 10.8611 +time/total (s) 2349.9 +Epoch -775 +---------------------------------- --------------- +2022-05-10 13:49:56.523080 PDT | [0] Epoch -774 finished +---------------------------------- --------------- +epoch -774 +replay_buffer/size 999033 +trainer/num train calls 227000 +trainer/Policy Loss -20.1359 +trainer/Log Pis Mean 23.6325 +trainer/Log Pis Std 12.9374 +trainer/Log Pis Max 63.5556 +trainer/Log Pis Min -14.6148 +trainer/policy/mean Mean -0.0386493 +trainer/policy/mean Std 0.901966 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82495 +trainer/policy/normal/std Std 0.605748 +trainer/policy/normal/std Max 5.66165 +trainer/policy/normal/std Min 0.384682 +trainer/policy/normal/log_std Mean 1.00564 +trainer/policy/normal/log_std Std 0.287839 +trainer/policy/normal/log_std Max 1.73372 +trainer/policy/normal/log_std Min -0.955338 +eval/num steps total 226154 +eval/num paths total 227 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0782648 +eval/Actions Std 0.90365 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55301 +time/logging (s) 0.00413369 +time/sampling batch (s) 0.286567 +time/saving (s) 0.00380227 +time/training (s) 6.64598 +time/epoch (s) 9.49349 +time/total (s) 2359.4 +Epoch -774 +---------------------------------- --------------- +2022-05-10 13:50:06.145856 PDT | [0] Epoch -773 finished +---------------------------------- --------------- +epoch -773 +replay_buffer/size 999033 +trainer/num train calls 228000 +trainer/Policy Loss -21.1756 +trainer/Log Pis Mean 24.5377 +trainer/Log Pis Std 13.8429 +trainer/Log Pis Max 75.6008 +trainer/Log Pis Min -7.31468 +trainer/policy/mean Mean -0.0538489 +trainer/policy/mean Std 0.907305 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.88685 +trainer/policy/normal/std Std 0.620789 +trainer/policy/normal/std Max 5.64113 +trainer/policy/normal/std Min 0.370527 +trainer/policy/normal/log_std Mean 1.0279 +trainer/policy/normal/log_std Std 0.282949 +trainer/policy/normal/log_std Max 1.73008 +trainer/policy/normal/log_std Min -0.992828 +eval/num steps total 226790 +eval/num paths total 228 +eval/path length Mean 636 +eval/path length Std 0 +eval/path length Max 636 +eval/path length Min 636 +eval/Rewards Mean 0.00157233 +eval/Rewards Std 0.0396214 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0287692 +eval/Actions Std 0.905943 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.7704 +time/logging (s) 0.00271677 +time/sampling batch (s) 0.29179 +time/saving (s) 0.00350203 +time/training (s) 6.53208 +time/epoch (s) 9.60049 +time/total (s) 2369 +Epoch -773 +---------------------------------- --------------- +2022-05-10 13:50:16.144448 PDT | [0] Epoch -772 finished +---------------------------------- --------------- +epoch -772 +replay_buffer/size 999033 +trainer/num train calls 229000 +trainer/Policy Loss -19.4427 +trainer/Log Pis Mean 24.8061 +trainer/Log Pis Std 12.725 +trainer/Log Pis Max 67.799 +trainer/Log Pis Min -6.06398 +trainer/policy/mean Mean -0.0233884 +trainer/policy/mean Std 0.906705 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80549 +trainer/policy/normal/std Std 0.604889 +trainer/policy/normal/std Max 5.78914 +trainer/policy/normal/std Min 0.384188 +trainer/policy/normal/log_std Mean 0.998771 +trainer/policy/normal/log_std Std 0.286493 +trainer/policy/normal/log_std Max 1.75598 +trainer/policy/normal/log_std Min -0.956623 +eval/num steps total 227790 +eval/num paths total 229 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.052899 +eval/Actions Std 0.914081 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58386 +time/logging (s) 0.00375809 +time/sampling batch (s) 0.284468 +time/saving (s) 0.00348885 +time/training (s) 7.10409 +time/epoch (s) 9.97967 +time/total (s) 2378.98 +Epoch -772 +---------------------------------- --------------- +2022-05-10 13:50:26.438203 PDT | [0] Epoch -771 finished +---------------------------------- --------------- +epoch -771 +replay_buffer/size 999033 +trainer/num train calls 230000 +trainer/Policy Loss -19.4061 +trainer/Log Pis Mean 24.526 +trainer/Log Pis Std 12.8542 +trainer/Log Pis Max 69.9809 +trainer/Log Pis Min -7.85077 +trainer/policy/mean Mean -0.0329728 +trainer/policy/mean Std 0.904147 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84163 +trainer/policy/normal/std Std 0.60621 +trainer/policy/normal/std Max 6.71696 +trainer/policy/normal/std Min 0.389533 +trainer/policy/normal/log_std Mean 1.01262 +trainer/policy/normal/log_std Std 0.28097 +trainer/policy/normal/log_std Max 1.90464 +trainer/policy/normal/log_std Min -0.942806 +eval/num steps total 228790 +eval/num paths total 230 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.472285 +eval/Actions Std 0.873906 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83371 +time/logging (s) 0.00368406 +time/sampling batch (s) 0.280182 +time/saving (s) 0.00349711 +time/training (s) 7.15284 +time/epoch (s) 10.2739 +time/total (s) 2389.26 +Epoch -771 +---------------------------------- --------------- +2022-05-10 13:50:36.288961 PDT | [0] Epoch -770 finished +---------------------------------- --------------- +epoch -770 +replay_buffer/size 999033 +trainer/num train calls 231000 +trainer/Policy Loss -19.674 +trainer/Log Pis Mean 24.1673 +trainer/Log Pis Std 12.7123 +trainer/Log Pis Max 66.2776 +trainer/Log Pis Min -6.50099 +trainer/policy/mean Mean -0.0266718 +trainer/policy/mean Std 0.907186 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85819 +trainer/policy/normal/std Std 0.612976 +trainer/policy/normal/std Max 5.83845 +trainer/policy/normal/std Min 0.386852 +trainer/policy/normal/log_std Mean 1.01797 +trainer/policy/normal/log_std Std 0.283745 +trainer/policy/normal/log_std Max 1.76446 +trainer/policy/normal/log_std Min -0.949714 +eval/num steps total 229790 +eval/num paths total 231 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.216191 +eval/Actions Std 0.842456 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6431 +time/logging (s) 0.00403046 +time/sampling batch (s) 0.281665 +time/saving (s) 0.00445065 +time/training (s) 6.89774 +time/epoch (s) 9.83099 +time/total (s) 2399.09 +Epoch -770 +---------------------------------- --------------- +2022-05-10 13:50:45.442365 PDT | [0] Epoch -769 finished +---------------------------------- --------------- +epoch -769 +replay_buffer/size 999033 +trainer/num train calls 232000 +trainer/Policy Loss -19.0206 +trainer/Log Pis Mean 24.6618 +trainer/Log Pis Std 13.355 +trainer/Log Pis Max 78.0788 +trainer/Log Pis Min -9.88202 +trainer/policy/mean Mean -0.0415898 +trainer/policy/mean Std 0.906408 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.74731 +trainer/policy/normal/std Std 0.582929 +trainer/policy/normal/std Max 5.43223 +trainer/policy/normal/std Min 0.340521 +trainer/policy/normal/log_std Mean 0.978573 +trainer/policy/normal/log_std Std 0.284067 +trainer/policy/normal/log_std Max 1.69235 +trainer/policy/normal/log_std Min -1.07728 +eval/num steps total 230790 +eval/num paths total 232 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111651 +eval/Actions Std 0.912679 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86201 +time/logging (s) 0.00395025 +time/sampling batch (s) 0.282439 +time/saving (s) 0.00402141 +time/training (s) 5.98054 +time/epoch (s) 9.13296 +time/total (s) 2408.23 +Epoch -769 +---------------------------------- --------------- +2022-05-10 13:50:56.423555 PDT | [0] Epoch -768 finished +---------------------------------- --------------- +epoch -768 +replay_buffer/size 999033 +trainer/num train calls 233000 +trainer/Policy Loss -20.2117 +trainer/Log Pis Mean 24.0341 +trainer/Log Pis Std 13.1799 +trainer/Log Pis Max 66.7631 +trainer/Log Pis Min -5.56148 +trainer/policy/mean Mean -0.0357925 +trainer/policy/mean Std 0.904635 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86567 +trainer/policy/normal/std Std 0.609638 +trainer/policy/normal/std Max 5.3764 +trainer/policy/normal/std Min 0.421106 +trainer/policy/normal/log_std Mean 1.0207 +trainer/policy/normal/log_std Std 0.283882 +trainer/policy/normal/log_std Max 1.68202 +trainer/policy/normal/log_std Min -0.86487 +eval/num steps total 231790 +eval/num paths total 233 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178992 +eval/Actions Std 0.90429 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7615 +time/logging (s) 0.00392703 +time/sampling batch (s) 0.533589 +time/saving (s) 0.00357815 +time/training (s) 7.65823 +time/epoch (s) 10.9608 +time/total (s) 2419.19 +Epoch -768 +---------------------------------- --------------- +2022-05-10 13:51:05.667897 PDT | [0] Epoch -767 finished +---------------------------------- --------------- +epoch -767 +replay_buffer/size 999033 +trainer/num train calls 234000 +trainer/Policy Loss -20.5519 +trainer/Log Pis Mean 24.1701 +trainer/Log Pis Std 13.1939 +trainer/Log Pis Max 69.0941 +trainer/Log Pis Min -8.67368 +trainer/policy/mean Mean -0.0343726 +trainer/policy/mean Std 0.909128 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84054 +trainer/policy/normal/std Std 0.584807 +trainer/policy/normal/std Max 5.94889 +trainer/policy/normal/std Min 0.381451 +trainer/policy/normal/log_std Mean 1.01443 +trainer/policy/normal/log_std Std 0.271419 +trainer/policy/normal/log_std Max 1.78321 +trainer/policy/normal/log_std Min -0.963774 +eval/num steps total 232790 +eval/num paths total 234 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0903931 +eval/Actions Std 0.915981 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5926 +time/logging (s) 0.00379942 +time/sampling batch (s) 0.280729 +time/saving (s) 0.00344977 +time/training (s) 6.34358 +time/epoch (s) 9.22416 +time/total (s) 2428.42 +Epoch -767 +---------------------------------- --------------- +2022-05-10 13:51:15.488179 PDT | [0] Epoch -766 finished +---------------------------------- --------------- +epoch -766 +replay_buffer/size 999033 +trainer/num train calls 235000 +trainer/Policy Loss -18.1213 +trainer/Log Pis Mean 25.3034 +trainer/Log Pis Std 12.9195 +trainer/Log Pis Max 75.2285 +trainer/Log Pis Min -8.98705 +trainer/policy/mean Mean -0.0419242 +trainer/policy/mean Std 0.899585 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84799 +trainer/policy/normal/std Std 0.645962 +trainer/policy/normal/std Max 5.69683 +trainer/policy/normal/std Min 0.416146 +trainer/policy/normal/log_std Mean 1.01053 +trainer/policy/normal/log_std Std 0.299773 +trainer/policy/normal/log_std Max 1.73991 +trainer/policy/normal/log_std Min -0.876719 +eval/num steps total 233790 +eval/num paths total 235 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0925736 +eval/Actions Std 0.912028 +eval/Actions Max 0.999998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74622 +time/logging (s) 0.0037002 +time/sampling batch (s) 0.28179 +time/saving (s) 0.00350306 +time/training (s) 6.76503 +time/epoch (s) 9.80025 +time/total (s) 2438.22 +Epoch -766 +---------------------------------- --------------- +2022-05-10 13:51:25.470782 PDT | [0] Epoch -765 finished +---------------------------------- --------------- +epoch -765 +replay_buffer/size 999033 +trainer/num train calls 236000 +trainer/Policy Loss -19.2457 +trainer/Log Pis Mean 25.1926 +trainer/Log Pis Std 13.0815 +trainer/Log Pis Max 78.3545 +trainer/Log Pis Min -4.27946 +trainer/policy/mean Mean -0.0448268 +trainer/policy/mean Std 0.904674 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86074 +trainer/policy/normal/std Std 0.627341 +trainer/policy/normal/std Max 5.2239 +trainer/policy/normal/std Min 0.368412 +trainer/policy/normal/log_std Mean 1.0163 +trainer/policy/normal/log_std Std 0.297459 +trainer/policy/normal/log_std Max 1.65324 +trainer/policy/normal/log_std Min -0.998553 +eval/num steps total 234790 +eval/num paths total 236 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178425 +eval/Actions Std 0.951233 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6782 +time/logging (s) 0.00374365 +time/sampling batch (s) 0.283669 +time/saving (s) 0.00399028 +time/training (s) 6.99302 +time/epoch (s) 9.96263 +time/total (s) 2448.19 +Epoch -765 +---------------------------------- --------------- +2022-05-10 13:51:35.504202 PDT | [0] Epoch -764 finished +---------------------------------- --------------- +epoch -764 +replay_buffer/size 999033 +trainer/num train calls 237000 +trainer/Policy Loss -19.662 +trainer/Log Pis Mean 24.3301 +trainer/Log Pis Std 12.657 +trainer/Log Pis Max 69.6871 +trainer/Log Pis Min -3.23063 +trainer/policy/mean Mean -0.0172708 +trainer/policy/mean Std 0.902132 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82969 +trainer/policy/normal/std Std 0.603036 +trainer/policy/normal/std Max 5.48946 +trainer/policy/normal/std Min 0.439485 +trainer/policy/normal/log_std Mean 1.00804 +trainer/policy/normal/log_std Std 0.28314 +trainer/policy/normal/log_std Max 1.70283 +trainer/policy/normal/log_std Min -0.822151 +eval/num steps total 235790 +eval/num paths total 237 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.198614 +eval/Actions Std 0.883951 +eval/Actions Max 0.999963 +eval/Actions Min -0.99997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49754 +time/logging (s) 0.00424396 +time/sampling batch (s) 0.281745 +time/saving (s) 0.00429951 +time/training (s) 7.22552 +time/epoch (s) 10.0134 +time/total (s) 2458.2 +Epoch -764 +---------------------------------- --------------- +2022-05-10 13:51:45.389934 PDT | [0] Epoch -763 finished +---------------------------------- --------------- +epoch -763 +replay_buffer/size 999033 +trainer/num train calls 238000 +trainer/Policy Loss -19.1885 +trainer/Log Pis Mean 24.3816 +trainer/Log Pis Std 13.8277 +trainer/Log Pis Max 62.6508 +trainer/Log Pis Min -8.7167 +trainer/policy/mean Mean -0.046892 +trainer/policy/mean Std 0.901745 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.91805 +trainer/policy/normal/std Std 0.578602 +trainer/policy/normal/std Max 6.32857 +trainer/policy/normal/std Min 0.419809 +trainer/policy/normal/log_std Mean 1.04505 +trainer/policy/normal/log_std Std 0.249395 +trainer/policy/normal/log_std Max 1.84507 +trainer/policy/normal/log_std Min -0.867956 +eval/num steps total 236790 +eval/num paths total 238 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0182288 +eval/Actions Std 0.902344 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63894 +time/logging (s) 0.00386331 +time/sampling batch (s) 0.281275 +time/saving (s) 0.00360091 +time/training (s) 6.93728 +time/epoch (s) 9.86496 +time/total (s) 2468.07 +Epoch -763 +---------------------------------- --------------- +2022-05-10 13:51:55.620229 PDT | [0] Epoch -762 finished +---------------------------------- --------------- +epoch -762 +replay_buffer/size 999033 +trainer/num train calls 239000 +trainer/Policy Loss -19.5633 +trainer/Log Pis Mean 23.151 +trainer/Log Pis Std 13.5562 +trainer/Log Pis Max 80.7388 +trainer/Log Pis Min -8.74705 +trainer/policy/mean Mean -0.0548684 +trainer/policy/mean Std 0.905719 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85207 +trainer/policy/normal/std Std 0.598646 +trainer/policy/normal/std Max 5.08927 +trainer/policy/normal/std Min 0.443699 +trainer/policy/normal/log_std Mean 1.01656 +trainer/policy/normal/log_std Std 0.281152 +trainer/policy/normal/log_std Max 1.62713 +trainer/policy/normal/log_std Min -0.81261 +eval/num steps total 237790 +eval/num paths total 239 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.287675 +eval/Actions Std 0.862937 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75135 +time/logging (s) 0.00382363 +time/sampling batch (s) 0.281336 +time/saving (s) 0.00355886 +time/training (s) 7.17018 +time/epoch (s) 10.2102 +time/total (s) 2478.29 +Epoch -762 +---------------------------------- --------------- +2022-05-10 13:52:05.554070 PDT | [0] Epoch -761 finished +---------------------------------- --------------- +epoch -761 +replay_buffer/size 999033 +trainer/num train calls 240000 +trainer/Policy Loss -18.0151 +trainer/Log Pis Mean 24.7041 +trainer/Log Pis Std 13.327 +trainer/Log Pis Max 69.0872 +trainer/Log Pis Min -13.4085 +trainer/policy/mean Mean -0.0219721 +trainer/policy/mean Std 0.90625 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86674 +trainer/policy/normal/std Std 0.624079 +trainer/policy/normal/std Max 5.58871 +trainer/policy/normal/std Min 0.344336 +trainer/policy/normal/log_std Mean 1.01965 +trainer/policy/normal/log_std Std 0.289633 +trainer/policy/normal/log_std Max 1.72075 +trainer/policy/normal/log_std Min -1.06614 +eval/num steps total 238790 +eval/num paths total 240 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.027945 +eval/Actions Std 0.909639 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52566 +time/logging (s) 0.00378973 +time/sampling batch (s) 0.531376 +time/saving (s) 0.003539 +time/training (s) 6.84941 +time/epoch (s) 9.91378 +time/total (s) 2488.2 +Epoch -761 +---------------------------------- --------------- +2022-05-10 13:52:15.623261 PDT | [0] Epoch -760 finished +---------------------------------- --------------- +epoch -760 +replay_buffer/size 999033 +trainer/num train calls 241000 +trainer/Policy Loss -18.0378 +trainer/Log Pis Mean 23.1089 +trainer/Log Pis Std 12.3476 +trainer/Log Pis Max 64.6883 +trainer/Log Pis Min -7.97177 +trainer/policy/mean Mean -0.0164602 +trainer/policy/mean Std 0.904131 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83751 +trainer/policy/normal/std Std 0.599548 +trainer/policy/normal/std Max 5.82933 +trainer/policy/normal/std Min 0.411538 +trainer/policy/normal/log_std Mean 1.01203 +trainer/policy/normal/log_std Std 0.27687 +trainer/policy/normal/log_std Max 1.7629 +trainer/policy/normal/log_std Min -0.887854 +eval/num steps total 239790 +eval/num paths total 241 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.321861 +eval/Actions Std 0.807563 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67397 +time/logging (s) 0.0037725 +time/sampling batch (s) 0.282121 +time/saving (s) 0.00352216 +time/training (s) 7.08563 +time/epoch (s) 10.049 +time/total (s) 2498.25 +Epoch -760 +---------------------------------- --------------- +2022-05-10 13:52:25.165247 PDT | [0] Epoch -759 finished +---------------------------------- --------------- +epoch -759 +replay_buffer/size 999033 +trainer/num train calls 242000 +trainer/Policy Loss -19.1789 +trainer/Log Pis Mean 24.9779 +trainer/Log Pis Std 13.5814 +trainer/Log Pis Max 71.4766 +trainer/Log Pis Min -6.93687 +trainer/policy/mean Mean -0.0503416 +trainer/policy/mean Std 0.904408 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85565 +trainer/policy/normal/std Std 0.576488 +trainer/policy/normal/std Max 5.20805 +trainer/policy/normal/std Min 0.440322 +trainer/policy/normal/log_std Mean 1.02117 +trainer/policy/normal/log_std Std 0.263277 +trainer/policy/normal/log_std Max 1.65021 +trainer/policy/normal/log_std Min -0.820249 +eval/num steps total 240197 +eval/num paths total 242 +eval/path length Mean 407 +eval/path length Std 0 +eval/path length Max 407 +eval/path length Min 407 +eval/Rewards Mean 0.002457 +eval/Rewards Std 0.0495072 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0184083 +eval/Actions Std 0.906208 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.61998 +time/logging (s) 0.00214384 +time/sampling batch (s) 0.282133 +time/saving (s) 0.0037848 +time/training (s) 6.61217 +time/epoch (s) 9.52021 +time/total (s) 2507.78 +Epoch -759 +---------------------------------- --------------- +2022-05-10 13:52:35.405256 PDT | [0] Epoch -758 finished +---------------------------------- --------------- +epoch -758 +replay_buffer/size 999033 +trainer/num train calls 243000 +trainer/Policy Loss -20.4201 +trainer/Log Pis Mean 25.3028 +trainer/Log Pis Std 12.7381 +trainer/Log Pis Max 85.0524 +trainer/Log Pis Min -2.48658 +trainer/policy/mean Mean -0.0332781 +trainer/policy/mean Std 0.910076 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85112 +trainer/policy/normal/std Std 0.584718 +trainer/policy/normal/std Max 4.75257 +trainer/policy/normal/std Min 0.372008 +trainer/policy/normal/log_std Mean 1.01716 +trainer/policy/normal/log_std Std 0.278715 +trainer/policy/normal/log_std Max 1.55869 +trainer/policy/normal/log_std Min -0.98884 +eval/num steps total 241197 +eval/num paths total 243 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00784712 +eval/Actions Std 0.955008 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81806 +time/logging (s) 0.00412819 +time/sampling batch (s) 0.531533 +time/saving (s) 0.00423932 +time/training (s) 6.86409 +time/epoch (s) 10.2221 +time/total (s) 2518 +Epoch -758 +---------------------------------- --------------- +2022-05-10 13:52:45.740191 PDT | [0] Epoch -757 finished +---------------------------------- --------------- +epoch -757 +replay_buffer/size 999033 +trainer/num train calls 244000 +trainer/Policy Loss -18.7187 +trainer/Log Pis Mean 23.953 +trainer/Log Pis Std 12.8162 +trainer/Log Pis Max 62.3295 +trainer/Log Pis Min -7.8518 +trainer/policy/mean Mean -0.0193055 +trainer/policy/mean Std 0.906003 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79667 +trainer/policy/normal/std Std 0.592753 +trainer/policy/normal/std Max 5.16125 +trainer/policy/normal/std Min 0.321753 +trainer/policy/normal/log_std Mean 0.995759 +trainer/policy/normal/log_std Std 0.288107 +trainer/policy/normal/log_std Max 1.64118 +trainer/policy/normal/log_std Min -1.13397 +eval/num steps total 242197 +eval/num paths total 244 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.285382 +eval/Actions Std 0.889371 +eval/Actions Max 0.999987 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.92033 +time/logging (s) 0.00374404 +time/sampling batch (s) 0.281911 +time/saving (s) 0.00344743 +time/training (s) 7.1048 +time/epoch (s) 10.3142 +time/total (s) 2528.32 +Epoch -757 +---------------------------------- --------------- +2022-05-10 13:52:56.356952 PDT | [0] Epoch -756 finished +---------------------------------- --------------- +epoch -756 +replay_buffer/size 999033 +trainer/num train calls 245000 +trainer/Policy Loss -19.4875 +trainer/Log Pis Mean 24.4122 +trainer/Log Pis Std 13.1284 +trainer/Log Pis Max 66.8603 +trainer/Log Pis Min -6.70609 +trainer/policy/mean Mean -0.0330034 +trainer/policy/mean Std 0.904549 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.87362 +trainer/policy/normal/std Std 0.599403 +trainer/policy/normal/std Max 5.44006 +trainer/policy/normal/std Min 0.369306 +trainer/policy/normal/log_std Mean 1.02414 +trainer/policy/normal/log_std Std 0.284048 +trainer/policy/normal/log_std Max 1.69379 +trainer/policy/normal/log_std Min -0.996129 +eval/num steps total 243197 +eval/num paths total 245 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0595314 +eval/Actions Std 0.914219 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73581 +time/logging (s) 0.00376018 +time/sampling batch (s) 0.284071 +time/saving (s) 0.00343246 +time/training (s) 7.56939 +time/epoch (s) 10.5965 +time/total (s) 2538.92 +Epoch -756 +---------------------------------- --------------- +2022-05-10 13:53:06.370329 PDT | [0] Epoch -755 finished +---------------------------------- --------------- +epoch -755 +replay_buffer/size 999033 +trainer/num train calls 246000 +trainer/Policy Loss -20.1443 +trainer/Log Pis Mean 23.4294 +trainer/Log Pis Std 13.2426 +trainer/Log Pis Max 60.5507 +trainer/Log Pis Min -11.436 +trainer/policy/mean Mean -0.0233886 +trainer/policy/mean Std 0.903403 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82903 +trainer/policy/normal/std Std 0.611987 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.348029 +trainer/policy/normal/log_std Mean 1.00698 +trainer/policy/normal/log_std Std 0.288216 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.05547 +eval/num steps total 244197 +eval/num paths total 246 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.172838 +eval/Actions Std 0.861841 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68234 +time/logging (s) 0.00387864 +time/sampling batch (s) 0.285676 +time/saving (s) 0.00358501 +time/training (s) 7.01775 +time/epoch (s) 9.99323 +time/total (s) 2548.91 +Epoch -755 +---------------------------------- --------------- +2022-05-10 13:53:16.598802 PDT | [0] Epoch -754 finished +---------------------------------- --------------- +epoch -754 +replay_buffer/size 999033 +trainer/num train calls 247000 +trainer/Policy Loss -19.2449 +trainer/Log Pis Mean 23.8558 +trainer/Log Pis Std 13.4519 +trainer/Log Pis Max 64.783 +trainer/Log Pis Min -11.5593 +trainer/policy/mean Mean -0.0291593 +trainer/policy/mean Std 0.90847 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84714 +trainer/policy/normal/std Std 0.595582 +trainer/policy/normal/std Max 5.52628 +trainer/policy/normal/std Min 0.393541 +trainer/policy/normal/log_std Mean 1.01678 +trainer/policy/normal/log_std Std 0.26783 +trainer/policy/normal/log_std Max 1.70952 +trainer/policy/normal/log_std Min -0.93257 +eval/num steps total 245197 +eval/num paths total 247 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0616699 +eval/Actions Std 0.905352 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.91944 +time/logging (s) 0.00378453 +time/sampling batch (s) 0.2896 +time/saving (s) 0.00354783 +time/training (s) 6.9915 +time/epoch (s) 10.2079 +time/total (s) 2559.13 +Epoch -754 +---------------------------------- --------------- +2022-05-10 13:53:27.742287 PDT | [0] Epoch -753 finished +---------------------------------- --------------- +epoch -753 +replay_buffer/size 999033 +trainer/num train calls 248000 +trainer/Policy Loss -19.5744 +trainer/Log Pis Mean 24.3569 +trainer/Log Pis Std 13.4443 +trainer/Log Pis Max 71.9898 +trainer/Log Pis Min -6.20829 +trainer/policy/mean Mean -0.0246021 +trainer/policy/mean Std 0.904705 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.8304 +trainer/policy/normal/std Std 0.602472 +trainer/policy/normal/std Max 5.12877 +trainer/policy/normal/std Min 0.43025 +trainer/policy/normal/log_std Mean 1.00845 +trainer/policy/normal/log_std Std 0.28249 +trainer/policy/normal/log_std Max 1.63487 +trainer/policy/normal/log_std Min -0.84339 +eval/num steps total 245644 +eval/num paths total 248 +eval/path length Mean 447 +eval/path length Std 0 +eval/path length Max 447 +eval/path length Min 447 +eval/Rewards Mean 0.00223714 +eval/Rewards Std 0.0472454 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean -0.0202052 +eval/Actions Std 0.90769 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.64615 +time/logging (s) 0.00254682 +time/sampling batch (s) 0.280559 +time/saving (s) 0.00409411 +time/training (s) 8.18883 +time/epoch (s) 11.1222 +time/total (s) 2570.25 +Epoch -753 +---------------------------------- --------------- +2022-05-10 13:53:38.259172 PDT | [0] Epoch -752 finished +---------------------------------- --------------- +epoch -752 +replay_buffer/size 999033 +trainer/num train calls 249000 +trainer/Policy Loss -19.7994 +trainer/Log Pis Mean 23.9312 +trainer/Log Pis Std 13.3548 +trainer/Log Pis Max 84.7759 +trainer/Log Pis Min -4.9634 +trainer/policy/mean Mean -0.0348669 +trainer/policy/mean Std 0.905872 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.88327 +trainer/policy/normal/std Std 0.611663 +trainer/policy/normal/std Max 5.90366 +trainer/policy/normal/std Min 0.397256 +trainer/policy/normal/log_std Mean 1.02711 +trainer/policy/normal/log_std Std 0.282872 +trainer/policy/normal/log_std Max 1.77557 +trainer/policy/normal/log_std Min -0.923173 +eval/num steps total 246644 +eval/num paths total 249 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.155757 +eval/Actions Std 0.894874 +eval/Actions Max 0.999996 +eval/Actions Min -0.999954 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64978 +time/logging (s) 0.00377916 +time/sampling batch (s) 0.283069 +time/saving (s) 0.00367967 +time/training (s) 7.55761 +time/epoch (s) 10.4979 +time/total (s) 2580.75 +Epoch -752 +---------------------------------- --------------- +2022-05-10 13:53:47.291385 PDT | [0] Epoch -751 finished +---------------------------------- --------------- +epoch -751 +replay_buffer/size 999033 +trainer/num train calls 250000 +trainer/Policy Loss -19.4808 +trainer/Log Pis Mean 23.1781 +trainer/Log Pis Std 13.1346 +trainer/Log Pis Max 67.0303 +trainer/Log Pis Min -5.95565 +trainer/policy/mean Mean -0.0790474 +trainer/policy/mean Std 0.899337 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79681 +trainer/policy/normal/std Std 0.59092 +trainer/policy/normal/std Max 5.23901 +trainer/policy/normal/std Min 0.378727 +trainer/policy/normal/log_std Mean 0.997109 +trainer/policy/normal/log_std Std 0.279957 +trainer/policy/normal/log_std Max 1.65613 +trainer/policy/normal/log_std Min -0.970938 +eval/num steps total 247644 +eval/num paths total 250 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0432795 +eval/Actions Std 0.901724 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85243 +time/logging (s) 0.00372965 +time/sampling batch (s) 0.279725 +time/saving (s) 0.00348023 +time/training (s) 5.87278 +time/epoch (s) 9.01213 +time/total (s) 2589.77 +Epoch -751 +---------------------------------- --------------- +2022-05-10 13:53:57.629230 PDT | [0] Epoch -750 finished +---------------------------------- --------------- +epoch -750 +replay_buffer/size 999033 +trainer/num train calls 251000 +trainer/Policy Loss -20.6194 +trainer/Log Pis Mean 23.7824 +trainer/Log Pis Std 12.6745 +trainer/Log Pis Max 63.6194 +trainer/Log Pis Min -10.6161 +trainer/policy/mean Mean -0.0308426 +trainer/policy/mean Std 0.902992 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80562 +trainer/policy/normal/std Std 0.624833 +trainer/policy/normal/std Max 5.2525 +trainer/policy/normal/std Min 0.385063 +trainer/policy/normal/log_std Mean 0.995269 +trainer/policy/normal/log_std Std 0.305068 +trainer/policy/normal/log_std Max 1.6587 +trainer/policy/normal/log_std Min -0.954347 +eval/num steps total 248644 +eval/num paths total 251 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113183 +eval/Actions Std 0.877947 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49552 +time/logging (s) 0.00370851 +time/sampling batch (s) 0.287158 +time/saving (s) 0.00352732 +time/training (s) 7.52766 +time/epoch (s) 10.3176 +time/total (s) 2600.09 +Epoch -750 +---------------------------------- --------------- +2022-05-10 13:54:08.409539 PDT | [0] Epoch -749 finished +---------------------------------- --------------- +epoch -749 +replay_buffer/size 999033 +trainer/num train calls 252000 +trainer/Policy Loss -20.6382 +trainer/Log Pis Mean 25.6876 +trainer/Log Pis Std 13.0713 +trainer/Log Pis Max 76.9075 +trainer/Log Pis Min -7.32761 +trainer/policy/mean Mean -0.0424995 +trainer/policy/mean Std 0.907753 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.90092 +trainer/policy/normal/std Std 0.604427 +trainer/policy/normal/std Max 5.12068 +trainer/policy/normal/std Min 0.412249 +trainer/policy/normal/log_std Mean 1.0347 +trainer/policy/normal/log_std Std 0.274433 +trainer/policy/normal/log_std Max 1.63329 +trainer/policy/normal/log_std Min -0.886127 +eval/num steps total 249644 +eval/num paths total 252 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0567609 +eval/Actions Std 0.903259 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67181 +time/logging (s) 0.00395053 +time/sampling batch (s) 0.27816 +time/saving (s) 0.00341923 +time/training (s) 7.80351 +time/epoch (s) 10.7608 +time/total (s) 2610.85 +Epoch -749 +---------------------------------- --------------- +2022-05-10 13:54:18.165700 PDT | [0] Epoch -748 finished +---------------------------------- --------------- +epoch -748 +replay_buffer/size 999033 +trainer/num train calls 253000 +trainer/Policy Loss -19.5735 +trainer/Log Pis Mean 24.8411 +trainer/Log Pis Std 13.4387 +trainer/Log Pis Max 71.9235 +trainer/Log Pis Min -4.69867 +trainer/policy/mean Mean -0.0357263 +trainer/policy/mean Std 0.906599 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.88734 +trainer/policy/normal/std Std 0.602188 +trainer/policy/normal/std Max 6.20582 +trainer/policy/normal/std Min 0.386353 +trainer/policy/normal/log_std Mean 1.03081 +trainer/policy/normal/log_std Std 0.268449 +trainer/policy/normal/log_std Max 1.82549 +trainer/policy/normal/log_std Min -0.951003 +eval/num steps total 250644 +eval/num paths total 253 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.216445 +eval/Actions Std 0.882576 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73698 +time/logging (s) 0.0037429 +time/sampling batch (s) 0.279038 +time/saving (s) 0.00363436 +time/training (s) 6.71262 +time/epoch (s) 9.73602 +time/total (s) 2620.59 +Epoch -748 +---------------------------------- --------------- +2022-05-10 13:54:28.789329 PDT | [0] Epoch -747 finished +---------------------------------- --------------- +epoch -747 +replay_buffer/size 999033 +trainer/num train calls 254000 +trainer/Policy Loss -19.8973 +trainer/Log Pis Mean 24.6082 +trainer/Log Pis Std 14.2087 +trainer/Log Pis Max 79.4229 +trainer/Log Pis Min -11.193 +trainer/policy/mean Mean -0.047324 +trainer/policy/mean Std 0.903807 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8234 +trainer/policy/normal/std Std 0.61207 +trainer/policy/normal/std Max 5.54904 +trainer/policy/normal/std Min 0.399956 +trainer/policy/normal/log_std Mean 1.00548 +trainer/policy/normal/log_std Std 0.283282 +trainer/policy/normal/log_std Max 1.71363 +trainer/policy/normal/log_std Min -0.9164 +eval/num steps total 251644 +eval/num paths total 254 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00390363 +eval/Actions Std 0.954039 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67583 +time/logging (s) 0.00396226 +time/sampling batch (s) 0.531228 +time/saving (s) 0.00402983 +time/training (s) 7.38886 +time/epoch (s) 10.6039 +time/total (s) 2631.2 +Epoch -747 +---------------------------------- --------------- +2022-05-10 13:54:38.444626 PDT | [0] Epoch -746 finished +---------------------------------- --------------- +epoch -746 +replay_buffer/size 999033 +trainer/num train calls 255000 +trainer/Policy Loss -20.2487 +trainer/Log Pis Mean 26.281 +trainer/Log Pis Std 13.4268 +trainer/Log Pis Max 69.2136 +trainer/Log Pis Min -4.08176 +trainer/policy/mean Mean -0.0214611 +trainer/policy/mean Std 0.902639 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86644 +trainer/policy/normal/std Std 0.611571 +trainer/policy/normal/std Max 5.16328 +trainer/policy/normal/std Min 0.374373 +trainer/policy/normal/log_std Mean 1.0212 +trainer/policy/normal/log_std Std 0.281405 +trainer/policy/normal/log_std Max 1.64157 +trainer/policy/normal/log_std Min -0.982503 +eval/num steps total 252644 +eval/num paths total 255 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.265439 +eval/Actions Std 0.876011 +eval/Actions Max 0.99998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60101 +time/logging (s) 0.00379607 +time/sampling batch (s) 0.533006 +time/saving (s) 0.00350966 +time/training (s) 6.49363 +time/epoch (s) 9.63495 +time/total (s) 2640.83 +Epoch -746 +---------------------------------- --------------- +2022-05-10 13:54:48.303890 PDT | [0] Epoch -745 finished +---------------------------------- --------------- +epoch -745 +replay_buffer/size 999033 +trainer/num train calls 256000 +trainer/Policy Loss -20.6389 +trainer/Log Pis Mean 24.2022 +trainer/Log Pis Std 13.1377 +trainer/Log Pis Max 70.3679 +trainer/Log Pis Min -8.0626 +trainer/policy/mean Mean -0.0377715 +trainer/policy/mean Std 0.906928 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.7726 +trainer/policy/normal/std Std 0.592895 +trainer/policy/normal/std Max 5.16424 +trainer/policy/normal/std Min 0.290665 +trainer/policy/normal/log_std Mean 0.985505 +trainer/policy/normal/log_std Std 0.2982 +trainer/policy/normal/log_std Max 1.64176 +trainer/policy/normal/log_std Min -1.23558 +eval/num steps total 253644 +eval/num paths total 256 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123269 +eval/Actions Std 0.920716 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61243 +time/logging (s) 0.00378782 +time/sampling batch (s) 0.29289 +time/saving (s) 0.00358011 +time/training (s) 6.9258 +time/epoch (s) 9.83849 +time/total (s) 2650.68 +Epoch -745 +---------------------------------- --------------- +2022-05-10 13:54:58.417200 PDT | [0] Epoch -744 finished +---------------------------------- -------------- +epoch -744 +replay_buffer/size 999033 +trainer/num train calls 257000 +trainer/Policy Loss -19.7102 +trainer/Log Pis Mean 25.3074 +trainer/Log Pis Std 13.0277 +trainer/Log Pis Max 66.5657 +trainer/Log Pis Min -9.17349 +trainer/policy/mean Mean -0.0228417 +trainer/policy/mean Std 0.911167 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.83563 +trainer/policy/normal/std Std 0.583532 +trainer/policy/normal/std Max 5.22664 +trainer/policy/normal/std Min 0.384107 +trainer/policy/normal/log_std Mean 1.0128 +trainer/policy/normal/log_std Std 0.270288 +trainer/policy/normal/log_std Max 1.65377 +trainer/policy/normal/log_std Min -0.956835 +eval/num steps total 254644 +eval/num paths total 257 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0644474 +eval/Actions Std 0.885353 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7601 +time/logging (s) 0.0038135 +time/sampling batch (s) 0.291823 +time/saving (s) 0.0035715 +time/training (s) 7.03316 +time/epoch (s) 10.0925 +time/total (s) 2660.77 +Epoch -744 +---------------------------------- -------------- +2022-05-10 13:55:08.403784 PDT | [0] Epoch -743 finished +---------------------------------- --------------- +epoch -743 +replay_buffer/size 999033 +trainer/num train calls 258000 +trainer/Policy Loss -19.7778 +trainer/Log Pis Mean 23.7177 +trainer/Log Pis Std 12.957 +trainer/Log Pis Max 64.5024 +trainer/Log Pis Min -8.50108 +trainer/policy/mean Mean -0.0359255 +trainer/policy/mean Std 0.908821 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7901 +trainer/policy/normal/std Std 0.586627 +trainer/policy/normal/std Max 5.1175 +trainer/policy/normal/std Min 0.37315 +trainer/policy/normal/log_std Mean 0.994646 +trainer/policy/normal/log_std Std 0.281235 +trainer/policy/normal/log_std Max 1.63267 +trainer/policy/normal/log_std Min -0.985776 +eval/num steps total 255644 +eval/num paths total 258 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0279232 +eval/Actions Std 0.908462 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59513 +time/logging (s) 0.00377172 +time/sampling batch (s) 0.296816 +time/saving (s) 0.00361047 +time/training (s) 7.06615 +time/epoch (s) 9.96548 +time/total (s) 2670.74 +Epoch -743 +---------------------------------- --------------- +2022-05-10 13:55:18.004743 PDT | [0] Epoch -742 finished +---------------------------------- --------------- +epoch -742 +replay_buffer/size 999033 +trainer/num train calls 259000 +trainer/Policy Loss -19.9163 +trainer/Log Pis Mean 23.839 +trainer/Log Pis Std 13.4796 +trainer/Log Pis Max 67.492 +trainer/Log Pis Min -9.59876 +trainer/policy/mean Mean -0.0347396 +trainer/policy/mean Std 0.900716 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81937 +trainer/policy/normal/std Std 0.590278 +trainer/policy/normal/std Max 5.27567 +trainer/policy/normal/std Min 0.379217 +trainer/policy/normal/log_std Mean 1.00488 +trainer/policy/normal/log_std Std 0.282906 +trainer/policy/normal/log_std Max 1.66311 +trainer/policy/normal/log_std Min -0.969646 +eval/num steps total 256644 +eval/num paths total 259 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.144017 +eval/Actions Std 0.888909 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89315 +time/logging (s) 0.00414663 +time/sampling batch (s) 0.292423 +time/saving (s) 0.00430153 +time/training (s) 6.38641 +time/epoch (s) 9.58044 +time/total (s) 2680.32 +Epoch -742 +---------------------------------- --------------- +2022-05-10 13:55:27.808643 PDT | [0] Epoch -741 finished +---------------------------------- --------------- +epoch -741 +replay_buffer/size 999033 +trainer/num train calls 260000 +trainer/Policy Loss -18.6808 +trainer/Log Pis Mean 24.2074 +trainer/Log Pis Std 13.7439 +trainer/Log Pis Max 60.1897 +trainer/Log Pis Min -12.4615 +trainer/policy/mean Mean -0.0439365 +trainer/policy/mean Std 0.90532 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86323 +trainer/policy/normal/std Std 0.596602 +trainer/policy/normal/std Max 5.11968 +trainer/policy/normal/std Min 0.434424 +trainer/policy/normal/log_std Mean 1.02197 +trainer/policy/normal/log_std Std 0.272012 +trainer/policy/normal/log_std Max 1.63309 +trainer/policy/normal/log_std Min -0.833734 +eval/num steps total 257644 +eval/num paths total 260 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0594991 +eval/Actions Std 0.897481 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.93699 +time/logging (s) 0.00371303 +time/sampling batch (s) 0.283597 +time/saving (s) 0.00349189 +time/training (s) 6.55499 +time/epoch (s) 9.78279 +time/total (s) 2690.11 +Epoch -741 +---------------------------------- --------------- +2022-05-10 13:55:37.132008 PDT | [0] Epoch -740 finished +---------------------------------- --------------- +epoch -740 +replay_buffer/size 999033 +trainer/num train calls 261000 +trainer/Policy Loss -18.9515 +trainer/Log Pis Mean 24.644 +trainer/Log Pis Std 12.9268 +trainer/Log Pis Max 64.57 +trainer/Log Pis Min -7.56361 +trainer/policy/mean Mean -0.0269562 +trainer/policy/mean Std 0.903241 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82289 +trainer/policy/normal/std Std 0.632587 +trainer/policy/normal/std Max 6.23229 +trainer/policy/normal/std Min 0.331868 +trainer/policy/normal/log_std Mean 1.00127 +trainer/policy/normal/log_std Std 0.304853 +trainer/policy/normal/log_std Max 1.82974 +trainer/policy/normal/log_std Min -1.10302 +eval/num steps total 258644 +eval/num paths total 261 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.160264 +eval/Actions Std 0.893791 +eval/Actions Max 0.999991 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78286 +time/logging (s) 0.00367056 +time/sampling batch (s) 0.280972 +time/saving (s) 0.00356478 +time/training (s) 6.23229 +time/epoch (s) 9.30335 +time/total (s) 2699.41 +Epoch -740 +---------------------------------- --------------- +2022-05-10 13:55:46.753645 PDT | [0] Epoch -739 finished +---------------------------------- --------------- +epoch -739 +replay_buffer/size 999033 +trainer/num train calls 262000 +trainer/Policy Loss -19.3975 +trainer/Log Pis Mean 25.3175 +trainer/Log Pis Std 13.529 +trainer/Log Pis Max 75.9393 +trainer/Log Pis Min -12.7225 +trainer/policy/mean Mean -0.0154516 +trainer/policy/mean Std 0.903909 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8618 +trainer/policy/normal/std Std 0.620041 +trainer/policy/normal/std Max 5.71953 +trainer/policy/normal/std Min 0.373184 +trainer/policy/normal/log_std Mean 1.01756 +trainer/policy/normal/log_std Std 0.293132 +trainer/policy/normal/log_std Max 1.74389 +trainer/policy/normal/log_std Min -0.985684 +eval/num steps total 259644 +eval/num paths total 262 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179984 +eval/Actions Std 0.912111 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63849 +time/logging (s) 0.003764 +time/sampling batch (s) 0.282961 +time/saving (s) 0.00364185 +time/training (s) 6.67287 +time/epoch (s) 9.60172 +time/total (s) 2709.02 +Epoch -739 +---------------------------------- --------------- +2022-05-10 13:55:56.998712 PDT | [0] Epoch -738 finished +---------------------------------- --------------- +epoch -738 +replay_buffer/size 999033 +trainer/num train calls 263000 +trainer/Policy Loss -19.867 +trainer/Log Pis Mean 25.2904 +trainer/Log Pis Std 13.6287 +trainer/Log Pis Max 86.268 +trainer/Log Pis Min -9.36824 +trainer/policy/mean Mean -0.0197677 +trainer/policy/mean Std 0.911519 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80463 +trainer/policy/normal/std Std 0.587605 +trainer/policy/normal/std Max 5.37995 +trainer/policy/normal/std Min 0.356017 +trainer/policy/normal/log_std Mean 1.00059 +trainer/policy/normal/log_std Std 0.276523 +trainer/policy/normal/log_std Max 1.68268 +trainer/policy/normal/log_std Min -1.03278 +eval/num steps total 260644 +eval/num paths total 263 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124777 +eval/Actions Std 0.896792 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68304 +time/logging (s) 0.00379214 +time/sampling batch (s) 0.28343 +time/saving (s) 0.003496 +time/training (s) 7.25111 +time/epoch (s) 10.2249 +time/total (s) 2719.25 +Epoch -738 +---------------------------------- --------------- +2022-05-10 13:56:07.146946 PDT | [0] Epoch -737 finished +---------------------------------- --------------- +epoch -737 +replay_buffer/size 999033 +trainer/num train calls 264000 +trainer/Policy Loss -19.8032 +trainer/Log Pis Mean 25.7242 +trainer/Log Pis Std 12.9128 +trainer/Log Pis Max 66.0569 +trainer/Log Pis Min -4.23296 +trainer/policy/mean Mean -0.0425539 +trainer/policy/mean Std 0.907126 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85631 +trainer/policy/normal/std Std 0.60105 +trainer/policy/normal/std Max 5.34567 +trainer/policy/normal/std Min 0.379792 +trainer/policy/normal/log_std Mean 1.01733 +trainer/policy/normal/log_std Std 0.285673 +trainer/policy/normal/log_std Max 1.67629 +trainer/policy/normal/log_std Min -0.968131 +eval/num steps total 261644 +eval/num paths total 264 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0639531 +eval/Actions Std 0.850905 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69985 +time/logging (s) 0.00420766 +time/sampling batch (s) 0.281381 +time/saving (s) 0.00422644 +time/training (s) 7.13893 +time/epoch (s) 10.1286 +time/total (s) 2729.38 +Epoch -737 +---------------------------------- --------------- +2022-05-10 13:56:17.724203 PDT | [0] Epoch -736 finished +---------------------------------- --------------- +epoch -736 +replay_buffer/size 999033 +trainer/num train calls 265000 +trainer/Policy Loss -18.6286 +trainer/Log Pis Mean 24.0774 +trainer/Log Pis Std 13.7221 +trainer/Log Pis Max 80.1166 +trainer/Log Pis Min -6.95828 +trainer/policy/mean Mean -0.0484172 +trainer/policy/mean Std 0.905877 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.80964 +trainer/policy/normal/std Std 0.60522 +trainer/policy/normal/std Max 4.76627 +trainer/policy/normal/std Min 0.357578 +trainer/policy/normal/log_std Mean 0.999833 +trainer/policy/normal/log_std Std 0.289268 +trainer/policy/normal/log_std Max 1.56156 +trainer/policy/normal/log_std Min -1.0284 +eval/num steps total 262644 +eval/num paths total 265 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0605307 +eval/Actions Std 0.914763 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61702 +time/logging (s) 0.00376397 +time/sampling batch (s) 0.531812 +time/saving (s) 0.00372948 +time/training (s) 7.39977 +time/epoch (s) 10.5561 +time/total (s) 2739.94 +Epoch -736 +---------------------------------- --------------- +2022-05-10 13:56:28.153106 PDT | [0] Epoch -735 finished +---------------------------------- --------------- +epoch -735 +replay_buffer/size 999033 +trainer/num train calls 266000 +trainer/Policy Loss -17.8937 +trainer/Log Pis Mean 24.6293 +trainer/Log Pis Std 13.3038 +trainer/Log Pis Max 71.3665 +trainer/Log Pis Min -8.1688 +trainer/policy/mean Mean -0.0467929 +trainer/policy/mean Std 0.900363 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83484 +trainer/policy/normal/std Std 0.61605 +trainer/policy/normal/std Max 6.04533 +trainer/policy/normal/std Min 0.351932 +trainer/policy/normal/log_std Mean 1.00672 +trainer/policy/normal/log_std Std 0.302263 +trainer/policy/normal/log_std Max 1.79929 +trainer/policy/normal/log_std Min -1.04432 +eval/num steps total 263644 +eval/num paths total 266 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.343054 +eval/Actions Std 0.849063 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67837 +time/logging (s) 0.00371915 +time/sampling batch (s) 0.783273 +time/saving (s) 0.00348704 +time/training (s) 6.93969 +time/epoch (s) 10.4085 +time/total (s) 2750.35 +Epoch -735 +---------------------------------- --------------- +2022-05-10 13:56:38.501519 PDT | [0] Epoch -734 finished +---------------------------------- --------------- +epoch -734 +replay_buffer/size 999033 +trainer/num train calls 267000 +trainer/Policy Loss -20.561 +trainer/Log Pis Mean 23.5908 +trainer/Log Pis Std 13.2928 +trainer/Log Pis Max 64.8723 +trainer/Log Pis Min -9.33535 +trainer/policy/mean Mean -0.0403185 +trainer/policy/mean Std 0.907259 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.82921 +trainer/policy/normal/std Std 0.597954 +trainer/policy/normal/std Max 5.15028 +trainer/policy/normal/std Min 0.308739 +trainer/policy/normal/log_std Mean 1.00813 +trainer/policy/normal/log_std Std 0.283836 +trainer/policy/normal/log_std Max 1.63905 +trainer/policy/normal/log_std Min -1.17526 +eval/num steps total 264644 +eval/num paths total 267 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.210537 +eval/Actions Std 0.914802 +eval/Actions Max 0.999976 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63642 +time/logging (s) 0.00402385 +time/sampling batch (s) 0.531066 +time/saving (s) 0.00358588 +time/training (s) 7.15352 +time/epoch (s) 10.3286 +time/total (s) 2760.68 +Epoch -734 +---------------------------------- --------------- +2022-05-10 13:56:49.436263 PDT | [0] Epoch -733 finished +---------------------------------- --------------- +epoch -733 +replay_buffer/size 999033 +trainer/num train calls 268000 +trainer/Policy Loss -19.2084 +trainer/Log Pis Mean 25.3261 +trainer/Log Pis Std 13.6346 +trainer/Log Pis Max 69.1013 +trainer/Log Pis Min -9.0081 +trainer/policy/mean Mean -0.033954 +trainer/policy/mean Std 0.906232 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89131 +trainer/policy/normal/std Std 0.634389 +trainer/policy/normal/std Max 6.63324 +trainer/policy/normal/std Min 0.375055 +trainer/policy/normal/log_std Mean 1.02786 +trainer/policy/normal/log_std Std 0.291026 +trainer/policy/normal/log_std Max 1.89209 +trainer/policy/normal/log_std Min -0.980682 +eval/num steps total 265644 +eval/num paths total 268 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128951 +eval/Actions Std 0.902889 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58839 +time/logging (s) 0.00380986 +time/sampling batch (s) 0.538677 +time/saving (s) 0.00352802 +time/training (s) 7.77958 +time/epoch (s) 10.914 +time/total (s) 2771.6 +Epoch -733 +---------------------------------- --------------- +2022-05-10 13:56:59.743862 PDT | [0] Epoch -732 finished +---------------------------------- --------------- +epoch -732 +replay_buffer/size 999033 +trainer/num train calls 269000 +trainer/Policy Loss -18.795 +trainer/Log Pis Mean 24.0531 +trainer/Log Pis Std 14.3056 +trainer/Log Pis Max 70.9146 +trainer/Log Pis Min -7.66602 +trainer/policy/mean Mean -0.032727 +trainer/policy/mean Std 0.904854 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.87342 +trainer/policy/normal/std Std 0.625764 +trainer/policy/normal/std Max 5.48128 +trainer/policy/normal/std Min 0.363092 +trainer/policy/normal/log_std Mean 1.02133 +trainer/policy/normal/log_std Std 0.29451 +trainer/policy/normal/log_std Max 1.70134 +trainer/policy/normal/log_std Min -1.0131 +eval/num steps total 266644 +eval/num paths total 269 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.275337 +eval/Actions Std 0.907198 +eval/Actions Max 0.999977 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77241 +time/logging (s) 0.00422851 +time/sampling batch (s) 0.289714 +time/saving (s) 0.00399074 +time/training (s) 7.21714 +time/epoch (s) 10.2875 +time/total (s) 2781.89 +Epoch -732 +---------------------------------- --------------- +2022-05-10 13:57:08.903976 PDT | [0] Epoch -731 finished +---------------------------------- --------------- +epoch -731 +replay_buffer/size 999033 +trainer/num train calls 270000 +trainer/Policy Loss -20.1117 +trainer/Log Pis Mean 25.3978 +trainer/Log Pis Std 13.7523 +trainer/Log Pis Max 71.7931 +trainer/Log Pis Min -8.44219 +trainer/policy/mean Mean -0.0334046 +trainer/policy/mean Std 0.905363 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.92406 +trainer/policy/normal/std Std 0.610352 +trainer/policy/normal/std Max 5.9706 +trainer/policy/normal/std Min 0.375823 +trainer/policy/normal/log_std Mean 1.04255 +trainer/policy/normal/log_std Std 0.275476 +trainer/policy/normal/log_std Max 1.78685 +trainer/policy/normal/log_std Min -0.978636 +eval/num steps total 267644 +eval/num paths total 270 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00295091 +eval/Actions Std 0.982928 +eval/Actions Max 0.999989 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45297 +time/logging (s) 0.00381689 +time/sampling batch (s) 0.280239 +time/saving (s) 0.00361636 +time/training (s) 6.39882 +time/epoch (s) 9.13947 +time/total (s) 2791.03 +Epoch -731 +---------------------------------- --------------- +2022-05-10 13:57:18.309095 PDT | [0] Epoch -730 finished +---------------------------------- --------------- +epoch -730 +replay_buffer/size 999033 +trainer/num train calls 271000 +trainer/Policy Loss -20.3336 +trainer/Log Pis Mean 24.0585 +trainer/Log Pis Std 13.714 +trainer/Log Pis Max 77.5215 +trainer/Log Pis Min -7.08646 +trainer/policy/mean Mean -0.0732015 +trainer/policy/mean Std 0.904488 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.90857 +trainer/policy/normal/std Std 0.600196 +trainer/policy/normal/std Max 5.42724 +trainer/policy/normal/std Min 0.365933 +trainer/policy/normal/log_std Mean 1.03848 +trainer/policy/normal/log_std Std 0.26862 +trainer/policy/normal/log_std Max 1.69143 +trainer/policy/normal/log_std Min -1.00531 +eval/num steps total 268644 +eval/num paths total 271 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.242049 +eval/Actions Std 0.812296 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61563 +time/logging (s) 0.00372976 +time/sampling batch (s) 0.528999 +time/saving (s) 0.00356659 +time/training (s) 6.23321 +time/epoch (s) 9.38513 +time/total (s) 2800.42 +Epoch -730 +---------------------------------- --------------- +2022-05-10 13:57:28.116668 PDT | [0] Epoch -729 finished +---------------------------------- --------------- +epoch -729 +replay_buffer/size 999033 +trainer/num train calls 272000 +trainer/Policy Loss -20.1881 +trainer/Log Pis Mean 24.5644 +trainer/Log Pis Std 12.7435 +trainer/Log Pis Max 68.6653 +trainer/Log Pis Min -5.30597 +trainer/policy/mean Mean -0.0510159 +trainer/policy/mean Std 0.904887 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83641 +trainer/policy/normal/std Std 0.617074 +trainer/policy/normal/std Max 5.14042 +trainer/policy/normal/std Min 0.35635 +trainer/policy/normal/log_std Mean 1.00811 +trainer/policy/normal/log_std Std 0.295579 +trainer/policy/normal/log_std Max 1.63714 +trainer/policy/normal/log_std Min -1.03184 +eval/num steps total 269644 +eval/num paths total 272 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00539138 +eval/Actions Std 0.932126 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88211 +time/logging (s) 0.00387662 +time/sampling batch (s) 0.279775 +time/saving (s) 0.00354122 +time/training (s) 6.61828 +time/epoch (s) 9.78758 +time/total (s) 2810.21 +Epoch -729 +---------------------------------- --------------- +2022-05-10 13:57:37.726306 PDT | [0] Epoch -728 finished +---------------------------------- --------------- +epoch -728 +replay_buffer/size 999033 +trainer/num train calls 273000 +trainer/Policy Loss -19.5325 +trainer/Log Pis Mean 22.7955 +trainer/Log Pis Std 12.9446 +trainer/Log Pis Max 61.6187 +trainer/Log Pis Min -7.35207 +trainer/policy/mean Mean -0.0520773 +trainer/policy/mean Std 0.909473 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8435 +trainer/policy/normal/std Std 0.591065 +trainer/policy/normal/std Max 5.11116 +trainer/policy/normal/std Min 0.30968 +trainer/policy/normal/log_std Mean 1.0155 +trainer/policy/normal/log_std Std 0.269343 +trainer/policy/normal/log_std Max 1.63143 +trainer/policy/normal/log_std Min -1.17221 +eval/num steps total 270644 +eval/num paths total 273 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0245194 +eval/Actions Std 0.876427 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89887 +time/logging (s) 0.0036654 +time/sampling batch (s) 0.280058 +time/saving (s) 0.00343364 +time/training (s) 6.40308 +time/epoch (s) 9.58911 +time/total (s) 2819.8 +Epoch -728 +---------------------------------- --------------- +2022-05-10 13:57:48.326827 PDT | [0] Epoch -727 finished +---------------------------------- --------------- +epoch -727 +replay_buffer/size 999033 +trainer/num train calls 274000 +trainer/Policy Loss -20.3666 +trainer/Log Pis Mean 24.6517 +trainer/Log Pis Std 13.0854 +trainer/Log Pis Max 60.6327 +trainer/Log Pis Min -15.2569 +trainer/policy/mean Mean -0.0187328 +trainer/policy/mean Std 0.911161 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81699 +trainer/policy/normal/std Std 0.594924 +trainer/policy/normal/std Max 5.67314 +trainer/policy/normal/std Min 0.361408 +trainer/policy/normal/log_std Mean 1.00341 +trainer/policy/normal/log_std Std 0.286493 +trainer/policy/normal/log_std Max 1.73574 +trainer/policy/normal/log_std Min -1.01775 +eval/num steps total 271644 +eval/num paths total 274 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.220625 +eval/Actions Std 0.938799 +eval/Actions Max 0.999986 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77651 +time/logging (s) 0.00374693 +time/sampling batch (s) 0.282813 +time/saving (s) 0.00345957 +time/training (s) 7.51372 +time/epoch (s) 10.5802 +time/total (s) 2830.38 +Epoch -727 +---------------------------------- --------------- +2022-05-10 13:57:58.690993 PDT | [0] Epoch -726 finished +---------------------------------- --------------- +epoch -726 +replay_buffer/size 999033 +trainer/num train calls 275000 +trainer/Policy Loss -19.3588 +trainer/Log Pis Mean 24.7462 +trainer/Log Pis Std 13.4894 +trainer/Log Pis Max 76.122 +trainer/Log Pis Min -8.009 +trainer/policy/mean Mean -0.0446056 +trainer/policy/mean Std 0.902515 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81315 +trainer/policy/normal/std Std 0.6022 +trainer/policy/normal/std Max 6.38605 +trainer/policy/normal/std Min 0.32568 +trainer/policy/normal/log_std Mean 1.00128 +trainer/policy/normal/log_std Std 0.289281 +trainer/policy/normal/log_std Max 1.85412 +trainer/policy/normal/log_std Min -1.12184 +eval/num steps total 272644 +eval/num paths total 275 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130109 +eval/Actions Std 0.812879 +eval/Actions Max 0.999991 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84969 +time/logging (s) 0.00430792 +time/sampling batch (s) 0.53507 +time/saving (s) 0.00427475 +time/training (s) 6.95083 +time/epoch (s) 10.3442 +time/total (s) 2840.73 +Epoch -726 +---------------------------------- --------------- +2022-05-10 13:58:08.017067 PDT | [0] Epoch -725 finished +---------------------------------- --------------- +epoch -725 +replay_buffer/size 999033 +trainer/num train calls 276000 +trainer/Policy Loss -18.9396 +trainer/Log Pis Mean 25.5306 +trainer/Log Pis Std 13.3502 +trainer/Log Pis Max 66.2737 +trainer/Log Pis Min -9.66526 +trainer/policy/mean Mean -0.0385221 +trainer/policy/mean Std 0.903389 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8563 +trainer/policy/normal/std Std 0.628151 +trainer/policy/normal/std Max 5.60838 +trainer/policy/normal/std Min 0.406451 +trainer/policy/normal/log_std Mean 1.01457 +trainer/policy/normal/log_std Std 0.29815 +trainer/policy/normal/log_std Max 1.72426 +trainer/policy/normal/log_std Min -0.900292 +eval/num steps total 273644 +eval/num paths total 276 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100847 +eval/Actions Std 0.902536 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55865 +time/logging (s) 0.00393658 +time/sampling batch (s) 0.285172 +time/saving (s) 0.00360887 +time/training (s) 6.45323 +time/epoch (s) 9.3046 +time/total (s) 2850.04 +Epoch -725 +---------------------------------- --------------- +2022-05-10 13:58:18.204457 PDT | [0] Epoch -724 finished +---------------------------------- --------------- +epoch -724 +replay_buffer/size 999033 +trainer/num train calls 277000 +trainer/Policy Loss -19.4545 +trainer/Log Pis Mean 23.8331 +trainer/Log Pis Std 13.1313 +trainer/Log Pis Max 64.8504 +trainer/Log Pis Min -10.8257 +trainer/policy/mean Mean -0.0420923 +trainer/policy/mean Std 0.900769 +trainer/policy/mean Max 0.999972 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.71999 +trainer/policy/normal/std Std 0.607752 +trainer/policy/normal/std Max 4.55147 +trainer/policy/normal/std Min 0.356855 +trainer/policy/normal/log_std Mean 0.963881 +trainer/policy/normal/log_std Std 0.305449 +trainer/policy/normal/log_std Max 1.51545 +trainer/policy/normal/log_std Min -1.03043 +eval/num steps total 274644 +eval/num paths total 277 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0347941 +eval/Actions Std 0.900296 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69543 +time/logging (s) 0.00385619 +time/sampling batch (s) 0.288259 +time/saving (s) 0.00353874 +time/training (s) 7.1754 +time/epoch (s) 10.1665 +time/total (s) 2860.21 +Epoch -724 +---------------------------------- --------------- +2022-05-10 13:58:27.868837 PDT | [0] Epoch -723 finished +---------------------------------- --------------- +epoch -723 +replay_buffer/size 999033 +trainer/num train calls 278000 +trainer/Policy Loss -18.7877 +trainer/Log Pis Mean 24.6152 +trainer/Log Pis Std 13.148 +trainer/Log Pis Max 60.4893 +trainer/Log Pis Min -8.43149 +trainer/policy/mean Mean -0.0467746 +trainer/policy/mean Std 0.903889 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.92079 +trainer/policy/normal/std Std 0.622438 +trainer/policy/normal/std Max 6.25734 +trainer/policy/normal/std Min 0.332746 +trainer/policy/normal/log_std Mean 1.04114 +trainer/policy/normal/log_std Std 0.27439 +trainer/policy/normal/log_std Max 1.83376 +trainer/policy/normal/log_std Min -1.10037 +eval/num steps total 275644 +eval/num paths total 278 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115951 +eval/Actions Std 0.915388 +eval/Actions Max 0.999991 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38877 +time/logging (s) 0.00366855 +time/sampling batch (s) 0.533713 +time/saving (s) 0.00345515 +time/training (s) 6.71419 +time/epoch (s) 9.6438 +time/total (s) 2869.85 +Epoch -723 +---------------------------------- --------------- +2022-05-10 13:58:37.271027 PDT | [0] Epoch -722 finished +---------------------------------- --------------- +epoch -722 +replay_buffer/size 999033 +trainer/num train calls 279000 +trainer/Policy Loss -19.3149 +trainer/Log Pis Mean 24.4662 +trainer/Log Pis Std 13.7404 +trainer/Log Pis Max 66.4919 +trainer/Log Pis Min -9.86952 +trainer/policy/mean Mean -0.0398865 +trainer/policy/mean Std 0.903387 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.79263 +trainer/policy/normal/std Std 0.613795 +trainer/policy/normal/std Max 5.52074 +trainer/policy/normal/std Min 0.393057 +trainer/policy/normal/log_std Mean 0.993295 +trainer/policy/normal/log_std Std 0.289128 +trainer/policy/normal/log_std Max 1.70851 +trainer/policy/normal/log_std Min -0.933799 +eval/num steps total 276644 +eval/num paths total 279 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.239509 +eval/Actions Std 0.947152 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56226 +time/logging (s) 0.00377649 +time/sampling batch (s) 0.286409 +time/saving (s) 0.00355087 +time/training (s) 6.52577 +time/epoch (s) 9.38176 +time/total (s) 2879.24 +Epoch -722 +---------------------------------- --------------- +2022-05-10 13:58:47.299154 PDT | [0] Epoch -721 finished +---------------------------------- --------------- +epoch -721 +replay_buffer/size 999033 +trainer/num train calls 280000 +trainer/Policy Loss -18.6047 +trainer/Log Pis Mean 24.4503 +trainer/Log Pis Std 12.9404 +trainer/Log Pis Max 63.0509 +trainer/Log Pis Min -10.4083 +trainer/policy/mean Mean -0.0448931 +trainer/policy/mean Std 0.902961 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86862 +trainer/policy/normal/std Std 0.603972 +trainer/policy/normal/std Max 5.02458 +trainer/policy/normal/std Min 0.379127 +trainer/policy/normal/log_std Mean 1.02213 +trainer/policy/normal/log_std Std 0.283161 +trainer/policy/normal/log_std Max 1.61434 +trainer/policy/normal/log_std Min -0.969885 +eval/num steps total 277644 +eval/num paths total 280 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0678352 +eval/Actions Std 0.899445 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73057 +time/logging (s) 0.00385824 +time/sampling batch (s) 0.279298 +time/saving (s) 0.00366252 +time/training (s) 6.99087 +time/epoch (s) 10.0083 +time/total (s) 2889.25 +Epoch -721 +---------------------------------- --------------- +2022-05-10 13:58:59.000366 PDT | [0] Epoch -720 finished +---------------------------------- --------------- +epoch -720 +replay_buffer/size 999033 +trainer/num train calls 281000 +trainer/Policy Loss -19.2203 +trainer/Log Pis Mean 23.704 +trainer/Log Pis Std 13.2412 +trainer/Log Pis Max 66.5364 +trainer/Log Pis Min -5.90262 +trainer/policy/mean Mean -0.0308387 +trainer/policy/mean Std 0.905129 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81928 +trainer/policy/normal/std Std 0.617781 +trainer/policy/normal/std Max 5.20148 +trainer/policy/normal/std Min 0.368647 +trainer/policy/normal/log_std Mean 1.0021 +trainer/policy/normal/log_std Std 0.295212 +trainer/policy/normal/log_std Max 1.64894 +trainer/policy/normal/log_std Min -0.997916 +eval/num steps total 278644 +eval/num paths total 281 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103858 +eval/Actions Std 0.914467 +eval/Actions Max 0.999995 +eval/Actions Min -0.999976 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59579 +time/logging (s) 0.00409821 +time/sampling batch (s) 0.530776 +time/saving (s) 0.00405515 +time/training (s) 8.54651 +time/epoch (s) 11.6812 +time/total (s) 2900.93 +Epoch -720 +---------------------------------- --------------- +2022-05-10 13:59:08.853135 PDT | [0] Epoch -719 finished +---------------------------------- --------------- +epoch -719 +replay_buffer/size 999033 +trainer/num train calls 282000 +trainer/Policy Loss -18.6791 +trainer/Log Pis Mean 24.522 +trainer/Log Pis Std 13.4981 +trainer/Log Pis Max 73.5724 +trainer/Log Pis Min -9.07927 +trainer/policy/mean Mean -0.0346759 +trainer/policy/mean Std 0.903362 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81985 +trainer/policy/normal/std Std 0.601934 +trainer/policy/normal/std Max 5.2866 +trainer/policy/normal/std Min 0.394681 +trainer/policy/normal/log_std Mean 1.0057 +trainer/policy/normal/log_std Std 0.276004 +trainer/policy/normal/log_std Max 1.66518 +trainer/policy/normal/log_std Min -0.929678 +eval/num steps total 279644 +eval/num paths total 282 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0173513 +eval/Actions Std 0.917146 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5021 +time/logging (s) 0.00378991 +time/sampling batch (s) 0.278929 +time/saving (s) 0.00343841 +time/training (s) 7.0439 +time/epoch (s) 9.83217 +time/total (s) 2910.77 +Epoch -719 +---------------------------------- --------------- +2022-05-10 13:59:19.174137 PDT | [0] Epoch -718 finished +---------------------------------- --------------- +epoch -718 +replay_buffer/size 999033 +trainer/num train calls 283000 +trainer/Policy Loss -19.5038 +trainer/Log Pis Mean 24.727 +trainer/Log Pis Std 13.5818 +trainer/Log Pis Max 64.4938 +trainer/Log Pis Min -6.55784 +trainer/policy/mean Mean -0.00341322 +trainer/policy/mean Std 0.904898 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86527 +trainer/policy/normal/std Std 0.623966 +trainer/policy/normal/std Max 5.59307 +trainer/policy/normal/std Min 0.305075 +trainer/policy/normal/log_std Mean 1.01842 +trainer/policy/normal/log_std Std 0.295145 +trainer/policy/normal/log_std Max 1.72153 +trainer/policy/normal/log_std Min -1.1872 +eval/num steps total 280644 +eval/num paths total 283 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0212341 +eval/Actions Std 0.946248 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58423 +time/logging (s) 0.00372811 +time/sampling batch (s) 0.532517 +time/saving (s) 0.00348188 +time/training (s) 7.17674 +time/epoch (s) 10.3007 +time/total (s) 2921.07 +Epoch -718 +---------------------------------- --------------- +2022-05-10 13:59:29.446790 PDT | [0] Epoch -717 finished +---------------------------------- --------------- +epoch -717 +replay_buffer/size 999033 +trainer/num train calls 284000 +trainer/Policy Loss -18.5256 +trainer/Log Pis Mean 24.6677 +trainer/Log Pis Std 13.4236 +trainer/Log Pis Max 68.4089 +trainer/Log Pis Min -6.73885 +trainer/policy/mean Mean -0.0412674 +trainer/policy/mean Std 0.901426 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.87697 +trainer/policy/normal/std Std 0.631159 +trainer/policy/normal/std Max 5.43468 +trainer/policy/normal/std Min 0.389276 +trainer/policy/normal/log_std Mean 1.02195 +trainer/policy/normal/log_std Std 0.297108 +trainer/policy/normal/log_std Max 1.6928 +trainer/policy/normal/log_std Min -0.943466 +eval/num steps total 281644 +eval/num paths total 284 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0968952 +eval/Actions Std 0.898175 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71907 +time/logging (s) 0.00373849 +time/sampling batch (s) 0.281341 +time/saving (s) 0.0035328 +time/training (s) 7.24472 +time/epoch (s) 10.2524 +time/total (s) 2931.33 +Epoch -717 +---------------------------------- --------------- +2022-05-10 13:59:39.709146 PDT | [0] Epoch -716 finished +---------------------------------- --------------- +epoch -716 +replay_buffer/size 999033 +trainer/num train calls 285000 +trainer/Policy Loss -19.6904 +trainer/Log Pis Mean 25.3278 +trainer/Log Pis Std 13.3426 +trainer/Log Pis Max 67.211 +trainer/Log Pis Min -6.98036 +trainer/policy/mean Mean -0.0128382 +trainer/policy/mean Std 0.902914 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83928 +trainer/policy/normal/std Std 0.607584 +trainer/policy/normal/std Max 5.21158 +trainer/policy/normal/std Min 0.352263 +trainer/policy/normal/log_std Mean 1.01054 +trainer/policy/normal/log_std Std 0.289333 +trainer/policy/normal/log_std Max 1.65088 +trainer/policy/normal/log_std Min -1.04338 +eval/num steps total 282644 +eval/num paths total 285 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.330013 +eval/Actions Std 0.849699 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80237 +time/logging (s) 0.00382736 +time/sampling batch (s) 0.282053 +time/saving (s) 0.00349845 +time/training (s) 7.15051 +time/epoch (s) 10.2423 +time/total (s) 2941.57 +Epoch -716 +---------------------------------- --------------- +2022-05-10 13:59:50.650427 PDT | [0] Epoch -715 finished +---------------------------------- --------------- +epoch -715 +replay_buffer/size 999033 +trainer/num train calls 286000 +trainer/Policy Loss -19.8911 +trainer/Log Pis Mean 24.1128 +trainer/Log Pis Std 12.9547 +trainer/Log Pis Max 69.4185 +trainer/Log Pis Min -3.88747 +trainer/policy/mean Mean -0.0402245 +trainer/policy/mean Std 0.908796 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82883 +trainer/policy/normal/std Std 0.622151 +trainer/policy/normal/std Max 4.99358 +trainer/policy/normal/std Min 0.300146 +trainer/policy/normal/log_std Mean 1.00433 +trainer/policy/normal/log_std Std 0.301126 +trainer/policy/normal/log_std Max 1.60815 +trainer/policy/normal/log_std Min -1.20349 +eval/num steps total 283644 +eval/num paths total 286 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.370335 +eval/Actions Std 0.838543 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67405 +time/logging (s) 0.00417739 +time/sampling batch (s) 0.533411 +time/saving (s) 0.00402388 +time/training (s) 7.70569 +time/epoch (s) 10.9214 +time/total (s) 2952.5 +Epoch -715 +---------------------------------- --------------- +2022-05-10 14:00:01.512360 PDT | [0] Epoch -714 finished +---------------------------------- --------------- +epoch -714 +replay_buffer/size 999033 +trainer/num train calls 287000 +trainer/Policy Loss -19.2918 +trainer/Log Pis Mean 24.1417 +trainer/Log Pis Std 13.5941 +trainer/Log Pis Max 69.1307 +trainer/Log Pis Min -10.1466 +trainer/policy/mean Mean -0.0370058 +trainer/policy/mean Std 0.904985 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81617 +trainer/policy/normal/std Std 0.613918 +trainer/policy/normal/std Max 5.49599 +trainer/policy/normal/std Min 0.331396 +trainer/policy/normal/log_std Mean 1.00081 +trainer/policy/normal/log_std Std 0.29619 +trainer/policy/normal/log_std Max 1.70402 +trainer/policy/normal/log_std Min -1.10444 +eval/num steps total 284644 +eval/num paths total 287 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.086129 +eval/Actions Std 0.89604 +eval/Actions Max 0.999982 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79878 +time/logging (s) 0.0037737 +time/sampling batch (s) 0.531166 +time/saving (s) 0.00365169 +time/training (s) 7.50347 +time/epoch (s) 10.8408 +time/total (s) 2963.34 +Epoch -714 +---------------------------------- --------------- +2022-05-10 14:00:12.312291 PDT | [0] Epoch -713 finished +---------------------------------- --------------- +epoch -713 +replay_buffer/size 999033 +trainer/num train calls 288000 +trainer/Policy Loss -19.38 +trainer/Log Pis Mean 24.6345 +trainer/Log Pis Std 13.8339 +trainer/Log Pis Max 76.0917 +trainer/Log Pis Min -10.9215 +trainer/policy/mean Mean -0.0398121 +trainer/policy/mean Std 0.903134 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84737 +trainer/policy/normal/std Std 0.619963 +trainer/policy/normal/std Max 5.42865 +trainer/policy/normal/std Min 0.332026 +trainer/policy/normal/log_std Mean 1.01223 +trainer/policy/normal/log_std Std 0.294077 +trainer/policy/normal/log_std Max 1.69169 +trainer/policy/normal/log_std Min -1.10254 +eval/num steps total 285644 +eval/num paths total 288 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0523489 +eval/Actions Std 0.897833 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54421 +time/logging (s) 0.0037503 +time/sampling batch (s) 1.03139 +time/saving (s) 0.00344334 +time/training (s) 7.19678 +time/epoch (s) 10.7796 +time/total (s) 2974.12 +Epoch -713 +---------------------------------- --------------- +2022-05-10 14:00:22.357710 PDT | [0] Epoch -712 finished +---------------------------------- --------------- +epoch -712 +replay_buffer/size 999033 +trainer/num train calls 289000 +trainer/Policy Loss -18.8873 +trainer/Log Pis Mean 24.6284 +trainer/Log Pis Std 12.9955 +trainer/Log Pis Max 71.3363 +trainer/Log Pis Min -6.97262 +trainer/policy/mean Mean -0.0568907 +trainer/policy/mean Std 0.905377 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81032 +trainer/policy/normal/std Std 0.589201 +trainer/policy/normal/std Max 6.21686 +trainer/policy/normal/std Min 0.417357 +trainer/policy/normal/log_std Mean 1.00273 +trainer/policy/normal/log_std Std 0.275471 +trainer/policy/normal/log_std Max 1.82726 +trainer/policy/normal/log_std Min -0.873814 +eval/num steps total 286644 +eval/num paths total 289 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.242778 +eval/Actions Std 0.910108 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51318 +time/logging (s) 0.00379736 +time/sampling batch (s) 0.284157 +time/saving (s) 0.00344432 +time/training (s) 7.22049 +time/epoch (s) 10.0251 +time/total (s) 2984.15 +Epoch -712 +---------------------------------- --------------- +2022-05-10 14:00:32.726216 PDT | [0] Epoch -711 finished +---------------------------------- --------------- +epoch -711 +replay_buffer/size 999033 +trainer/num train calls 290000 +trainer/Policy Loss -20.3183 +trainer/Log Pis Mean 24.9199 +trainer/Log Pis Std 12.7714 +trainer/Log Pis Max 59.7907 +trainer/Log Pis Min -6.60729 +trainer/policy/mean Mean -0.0281503 +trainer/policy/mean Std 0.907352 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79069 +trainer/policy/normal/std Std 0.604051 +trainer/policy/normal/std Max 4.85279 +trainer/policy/normal/std Min 0.35585 +trainer/policy/normal/log_std Mean 0.99239 +trainer/policy/normal/log_std Std 0.292972 +trainer/policy/normal/log_std Max 1.57955 +trainer/policy/normal/log_std Min -1.03325 +eval/num steps total 287644 +eval/num paths total 290 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.451054 +eval/Actions Std 0.833382 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59653 +time/logging (s) 0.00366564 +time/sampling batch (s) 0.53201 +time/saving (s) 0.00349138 +time/training (s) 7.21242 +time/epoch (s) 10.3481 +time/total (s) 2994.5 +Epoch -711 +---------------------------------- --------------- +2022-05-10 14:00:42.047369 PDT | [0] Epoch -710 finished +---------------------------------- --------------- +epoch -710 +replay_buffer/size 999033 +trainer/num train calls 291000 +trainer/Policy Loss -19.4735 +trainer/Log Pis Mean 24.088 +trainer/Log Pis Std 12.9435 +trainer/Log Pis Max 66.5801 +trainer/Log Pis Min -8.72132 +trainer/policy/mean Mean -0.049909 +trainer/policy/mean Std 0.902948 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76766 +trainer/policy/normal/std Std 0.601554 +trainer/policy/normal/std Max 5.17036 +trainer/policy/normal/std Min 0.427135 +trainer/policy/normal/log_std Mean 0.984503 +trainer/policy/normal/log_std Std 0.289665 +trainer/policy/normal/log_std Max 1.64294 +trainer/policy/normal/log_std Min -0.850655 +eval/num steps total 288644 +eval/num paths total 291 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0823173 +eval/Actions Std 0.915579 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51325 +time/logging (s) 0.00419772 +time/sampling batch (s) 0.285984 +time/saving (s) 0.00415603 +time/training (s) 6.49362 +time/epoch (s) 9.30121 +time/total (s) 3003.81 +Epoch -710 +---------------------------------- --------------- +2022-05-10 14:00:52.935937 PDT | [0] Epoch -709 finished +---------------------------------- --------------- +epoch -709 +replay_buffer/size 999033 +trainer/num train calls 292000 +trainer/Policy Loss -19.9745 +trainer/Log Pis Mean 24.6629 +trainer/Log Pis Std 13.0297 +trainer/Log Pis Max 71.9392 +trainer/Log Pis Min -7.33232 +trainer/policy/mean Mean -0.0364758 +trainer/policy/mean Std 0.905477 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.88418 +trainer/policy/normal/std Std 0.599235 +trainer/policy/normal/std Max 5.71596 +trainer/policy/normal/std Min 0.444931 +trainer/policy/normal/log_std Mean 1.03032 +trainer/policy/normal/log_std Std 0.264651 +trainer/policy/normal/log_std Max 1.74326 +trainer/policy/normal/log_std Min -0.809836 +eval/num steps total 289644 +eval/num paths total 292 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.07201 +eval/Actions Std 0.91699 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63772 +time/logging (s) 0.00381381 +time/sampling batch (s) 0.282015 +time/saving (s) 0.0037397 +time/training (s) 7.94012 +time/epoch (s) 10.8674 +time/total (s) 3014.68 +Epoch -709 +---------------------------------- --------------- +2022-05-10 14:01:02.971091 PDT | [0] Epoch -708 finished +---------------------------------- --------------- +epoch -708 +replay_buffer/size 999033 +trainer/num train calls 293000 +trainer/Policy Loss -19.2873 +trainer/Log Pis Mean 24.6396 +trainer/Log Pis Std 12.7442 +trainer/Log Pis Max 73.5241 +trainer/Log Pis Min -6.83296 +trainer/policy/mean Mean -0.0172264 +trainer/policy/mean Std 0.905854 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.83794 +trainer/policy/normal/std Std 0.616518 +trainer/policy/normal/std Max 5.41898 +trainer/policy/normal/std Min 0.3451 +trainer/policy/normal/log_std Mean 1.00976 +trainer/policy/normal/log_std Std 0.288801 +trainer/policy/normal/log_std Max 1.68991 +trainer/policy/normal/log_std Min -1.06392 +eval/num steps total 290644 +eval/num paths total 293 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.204684 +eval/Actions Std 0.954131 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46796 +time/logging (s) 0.00379037 +time/sampling batch (s) 0.281749 +time/saving (s) 0.00348551 +time/training (s) 7.25764 +time/epoch (s) 10.0146 +time/total (s) 3024.7 +Epoch -708 +---------------------------------- --------------- +2022-05-10 14:01:13.275140 PDT | [0] Epoch -707 finished +---------------------------------- --------------- +epoch -707 +replay_buffer/size 999033 +trainer/num train calls 294000 +trainer/Policy Loss -19.6471 +trainer/Log Pis Mean 24.9087 +trainer/Log Pis Std 13.1083 +trainer/Log Pis Max 70.0095 +trainer/Log Pis Min -14.6286 +trainer/policy/mean Mean -0.0451597 +trainer/policy/mean Std 0.906278 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.76726 +trainer/policy/normal/std Std 0.61426 +trainer/policy/normal/std Max 6.28706 +trainer/policy/normal/std Min 0.32365 +trainer/policy/normal/log_std Mean 0.980716 +trainer/policy/normal/log_std Std 0.311676 +trainer/policy/normal/log_std Max 1.83849 +trainer/policy/normal/log_std Min -1.12809 +eval/num steps total 291644 +eval/num paths total 294 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113683 +eval/Actions Std 0.916177 +eval/Actions Max 0.999988 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56627 +time/logging (s) 0.00374414 +time/sampling batch (s) 0.531926 +time/saving (s) 0.00353514 +time/training (s) 7.17832 +time/epoch (s) 10.2838 +time/total (s) 3034.98 +Epoch -707 +---------------------------------- --------------- +2022-05-10 14:01:23.503214 PDT | [0] Epoch -706 finished +---------------------------------- --------------- +epoch -706 +replay_buffer/size 999033 +trainer/num train calls 295000 +trainer/Policy Loss -19.1832 +trainer/Log Pis Mean 24.3816 +trainer/Log Pis Std 13.0748 +trainer/Log Pis Max 73.599 +trainer/Log Pis Min -4.99313 +trainer/policy/mean Mean -0.0413106 +trainer/policy/mean Std 0.908861 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77342 +trainer/policy/normal/std Std 0.613773 +trainer/policy/normal/std Max 6.0963 +trainer/policy/normal/std Min 0.375508 +trainer/policy/normal/log_std Mean 0.984817 +trainer/policy/normal/log_std Std 0.298534 +trainer/policy/normal/log_std Max 1.80768 +trainer/policy/normal/log_std Min -0.979476 +eval/num steps total 292644 +eval/num paths total 295 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.328139 +eval/Actions Std 0.758081 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60867 +time/logging (s) 0.00372262 +time/sampling batch (s) 0.280191 +time/saving (s) 0.00352965 +time/training (s) 7.31185 +time/epoch (s) 10.208 +time/total (s) 3045.19 +Epoch -706 +---------------------------------- --------------- +2022-05-10 14:01:34.180276 PDT | [0] Epoch -705 finished +---------------------------------- --------------- +epoch -705 +replay_buffer/size 999033 +trainer/num train calls 296000 +trainer/Policy Loss -19.8811 +trainer/Log Pis Mean 23.9937 +trainer/Log Pis Std 13.6208 +trainer/Log Pis Max 68.0968 +trainer/Log Pis Min -10.1843 +trainer/policy/mean Mean -0.0575724 +trainer/policy/mean Std 0.905373 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81502 +trainer/policy/normal/std Std 0.603309 +trainer/policy/normal/std Max 7.20015 +trainer/policy/normal/std Min 0.386217 +trainer/policy/normal/log_std Mean 1.00353 +trainer/policy/normal/log_std Std 0.279125 +trainer/policy/normal/log_std Max 1.9741 +trainer/policy/normal/log_std Min -0.951355 +eval/num steps total 293644 +eval/num paths total 296 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108061 +eval/Actions Std 0.892808 +eval/Actions Max 0.99999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65438 +time/logging (s) 0.00365188 +time/sampling batch (s) 0.531105 +time/saving (s) 0.00340109 +time/training (s) 7.46422 +time/epoch (s) 10.6568 +time/total (s) 3055.85 +Epoch -705 +---------------------------------- --------------- +2022-05-10 14:01:43.985801 PDT | [0] Epoch -704 finished +---------------------------------- --------------- +epoch -704 +replay_buffer/size 999033 +trainer/num train calls 297000 +trainer/Policy Loss -18.926 +trainer/Log Pis Mean 24.1007 +trainer/Log Pis Std 12.5527 +trainer/Log Pis Max 72.0541 +trainer/Log Pis Min -4.56965 +trainer/policy/mean Mean -0.0424867 +trainer/policy/mean Std 0.907156 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999976 +trainer/policy/normal/std Mean 2.82488 +trainer/policy/normal/std Std 0.6183 +trainer/policy/normal/std Max 5.33352 +trainer/policy/normal/std Min 0.373501 +trainer/policy/normal/log_std Mean 1.00335 +trainer/policy/normal/log_std Std 0.299883 +trainer/policy/normal/log_std Max 1.67401 +trainer/policy/normal/log_std Min -0.984835 +eval/num steps total 294644 +eval/num paths total 297 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.434159 +eval/Actions Std 0.717008 +eval/Actions Max 0.999992 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39023 +time/logging (s) 0.00398968 +time/sampling batch (s) 0.285548 +time/saving (s) 0.00402124 +time/training (s) 7.10156 +time/epoch (s) 9.78534 +time/total (s) 3065.64 +Epoch -704 +---------------------------------- --------------- +2022-05-10 14:01:55.445386 PDT | [0] Epoch -703 finished +---------------------------------- --------------- +epoch -703 +replay_buffer/size 999033 +trainer/num train calls 298000 +trainer/Policy Loss -20.3659 +trainer/Log Pis Mean 24.4424 +trainer/Log Pis Std 12.594 +trainer/Log Pis Max 64.6801 +trainer/Log Pis Min -5.96033 +trainer/policy/mean Mean -0.0339844 +trainer/policy/mean Std 0.907469 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85778 +trainer/policy/normal/std Std 0.608086 +trainer/policy/normal/std Max 5.55175 +trainer/policy/normal/std Min 0.342433 +trainer/policy/normal/log_std Mean 1.01897 +trainer/policy/normal/log_std Std 0.276905 +trainer/policy/normal/log_std Max 1.71411 +trainer/policy/normal/log_std Min -1.07168 +eval/num steps total 295644 +eval/num paths total 298 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.235407 +eval/Actions Std 0.920184 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62553 +time/logging (s) 0.00379126 +time/sampling batch (s) 0.534194 +time/saving (s) 0.0036266 +time/training (s) 8.27153 +time/epoch (s) 11.4387 +time/total (s) 3077.08 +Epoch -703 +---------------------------------- --------------- +2022-05-10 14:02:05.149430 PDT | [0] Epoch -702 finished +---------------------------------- --------------- +epoch -702 +replay_buffer/size 999033 +trainer/num train calls 299000 +trainer/Policy Loss -20.1553 +trainer/Log Pis Mean 24.4138 +trainer/Log Pis Std 12.8428 +trainer/Log Pis Max 73.083 +trainer/Log Pis Min -5.88901 +trainer/policy/mean Mean -0.013548 +trainer/policy/mean Std 0.906309 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8059 +trainer/policy/normal/std Std 0.602495 +trainer/policy/normal/std Max 5.45243 +trainer/policy/normal/std Min 0.368375 +trainer/policy/normal/log_std Mean 0.997824 +trainer/policy/normal/log_std Std 0.295287 +trainer/policy/normal/log_std Max 1.69606 +trainer/policy/normal/log_std Min -0.998654 +eval/num steps total 296644 +eval/num paths total 299 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0333005 +eval/Actions Std 0.917607 +eval/Actions Max 0.999988 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60791 +time/logging (s) 0.00384045 +time/sampling batch (s) 0.281837 +time/saving (s) 0.00347184 +time/training (s) 6.78664 +time/epoch (s) 9.6837 +time/total (s) 3086.77 +Epoch -702 +---------------------------------- --------------- +2022-05-10 14:02:16.541829 PDT | [0] Epoch -701 finished +---------------------------------- --------------- +epoch -701 +replay_buffer/size 999033 +trainer/num train calls 300000 +trainer/Policy Loss -19.6731 +trainer/Log Pis Mean 24.7761 +trainer/Log Pis Std 13.0412 +trainer/Log Pis Max 66.9601 +trainer/Log Pis Min -7.96329 +trainer/policy/mean Mean -0.0515764 +trainer/policy/mean Std 0.898237 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82485 +trainer/policy/normal/std Std 0.631093 +trainer/policy/normal/std Max 5.08901 +trainer/policy/normal/std Min 0.380568 +trainer/policy/normal/log_std Mean 1.00191 +trainer/policy/normal/log_std Std 0.305724 +trainer/policy/normal/log_std Max 1.62708 +trainer/policy/normal/log_std Min -0.966091 +eval/num steps total 297644 +eval/num paths total 300 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119727 +eval/Actions Std 0.889119 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59239 +time/logging (s) 0.00374564 +time/sampling batch (s) 0.536494 +time/saving (s) 0.00640951 +time/training (s) 8.23271 +time/epoch (s) 11.3717 +time/total (s) 3098.14 +Epoch -701 +---------------------------------- --------------- +2022-05-10 14:02:27.416445 PDT | [0] Epoch -700 finished +---------------------------------- --------------- +epoch -700 +replay_buffer/size 999033 +trainer/num train calls 301000 +trainer/Policy Loss -19.9069 +trainer/Log Pis Mean 23.9525 +trainer/Log Pis Std 12.5543 +trainer/Log Pis Max 69.4248 +trainer/Log Pis Min -4.09903 +trainer/policy/mean Mean -0.0359356 +trainer/policy/mean Std 0.908347 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82206 +trainer/policy/normal/std Std 0.589791 +trainer/policy/normal/std Max 5.11694 +trainer/policy/normal/std Min 0.419931 +trainer/policy/normal/log_std Mean 1.00697 +trainer/policy/normal/log_std Std 0.275173 +trainer/policy/normal/log_std Max 1.63256 +trainer/policy/normal/log_std Min -0.867665 +eval/num steps total 298644 +eval/num paths total 301 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101348 +eval/Actions Std 0.895181 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60996 +time/logging (s) 0.00383966 +time/sampling batch (s) 0.533876 +time/saving (s) 0.00344628 +time/training (s) 7.70306 +time/epoch (s) 10.8542 +time/total (s) 3109 +Epoch -700 +---------------------------------- --------------- +2022-05-10 14:02:38.020151 PDT | [0] Epoch -699 finished +---------------------------------- -------------- +epoch -699 +replay_buffer/size 999033 +trainer/num train calls 302000 +trainer/Policy Loss -18.9244 +trainer/Log Pis Mean 24.1878 +trainer/Log Pis Std 12.9891 +trainer/Log Pis Max 78.7467 +trainer/Log Pis Min -9.97458 +trainer/policy/mean Mean -0.0490073 +trainer/policy/mean Std 0.904664 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85344 +trainer/policy/normal/std Std 0.609417 +trainer/policy/normal/std Max 5.25469 +trainer/policy/normal/std Min 0.366429 +trainer/policy/normal/log_std Mean 1.01589 +trainer/policy/normal/log_std Std 0.286735 +trainer/policy/normal/log_std Max 1.65912 +trainer/policy/normal/log_std Min -1.00395 +eval/num steps total 299644 +eval/num paths total 302 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0388875 +eval/Actions Std 0.946025 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63945 +time/logging (s) 0.0037559 +time/sampling batch (s) 0.283332 +time/saving (s) 0.0037753 +time/training (s) 7.65266 +time/epoch (s) 10.583 +time/total (s) 3119.59 +Epoch -699 +---------------------------------- -------------- +2022-05-10 14:02:48.038239 PDT | [0] Epoch -698 finished +---------------------------------- --------------- +epoch -698 +replay_buffer/size 999033 +trainer/num train calls 303000 +trainer/Policy Loss -18.2605 +trainer/Log Pis Mean 24.2081 +trainer/Log Pis Std 13.4868 +trainer/Log Pis Max 72.2813 +trainer/Log Pis Min -11.456 +trainer/policy/mean Mean -0.0190029 +trainer/policy/mean Std 0.905965 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.84211 +trainer/policy/normal/std Std 0.613674 +trainer/policy/normal/std Max 5.61425 +trainer/policy/normal/std Min 0.351474 +trainer/policy/normal/log_std Mean 1.01132 +trainer/policy/normal/log_std Std 0.290002 +trainer/policy/normal/log_std Max 1.72531 +trainer/policy/normal/log_std Min -1.04562 +eval/num steps total 300644 +eval/num paths total 303 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.424716 +eval/Actions Std 0.890103 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66661 +time/logging (s) 0.00404499 +time/sampling batch (s) 0.284648 +time/saving (s) 0.0039989 +time/training (s) 7.03816 +time/epoch (s) 9.99747 +time/total (s) 3129.59 +Epoch -698 +---------------------------------- --------------- +2022-05-10 14:02:58.108150 PDT | [0] Epoch -697 finished +---------------------------------- --------------- +epoch -697 +replay_buffer/size 999033 +trainer/num train calls 304000 +trainer/Policy Loss -19.3972 +trainer/Log Pis Mean 24.5151 +trainer/Log Pis Std 12.6726 +trainer/Log Pis Max 63.9856 +trainer/Log Pis Min -3.81596 +trainer/policy/mean Mean -0.0150425 +trainer/policy/mean Std 0.905449 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84709 +trainer/policy/normal/std Std 0.618934 +trainer/policy/normal/std Max 6.26055 +trainer/policy/normal/std Min 0.319143 +trainer/policy/normal/log_std Mean 1.01322 +trainer/policy/normal/log_std Std 0.287919 +trainer/policy/normal/log_std Max 1.83427 +trainer/policy/normal/log_std Min -1.14211 +eval/num steps total 301644 +eval/num paths total 304 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0998393 +eval/Actions Std 0.920394 +eval/Actions Max 0.999973 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59846 +time/logging (s) 0.00369598 +time/sampling batch (s) 0.284406 +time/saving (s) 0.0035092 +time/training (s) 7.1585 +time/epoch (s) 10.0486 +time/total (s) 3139.64 +Epoch -697 +---------------------------------- --------------- +2022-05-10 14:03:08.328992 PDT | [0] Epoch -696 finished +---------------------------------- --------------- +epoch -696 +replay_buffer/size 999033 +trainer/num train calls 305000 +trainer/Policy Loss -18.8406 +trainer/Log Pis Mean 22.6934 +trainer/Log Pis Std 12.5873 +trainer/Log Pis Max 62.5323 +trainer/Log Pis Min -12.6807 +trainer/policy/mean Mean -0.0305342 +trainer/policy/mean Std 0.906088 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77647 +trainer/policy/normal/std Std 0.596704 +trainer/policy/normal/std Max 5.42717 +trainer/policy/normal/std Min 0.386669 +trainer/policy/normal/log_std Mean 0.988394 +trainer/policy/normal/log_std Std 0.28742 +trainer/policy/normal/log_std Max 1.69142 +trainer/policy/normal/log_std Min -0.950187 +eval/num steps total 302644 +eval/num paths total 305 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00473624 +eval/Actions Std 0.902043 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62498 +time/logging (s) 0.0042384 +time/sampling batch (s) 0.283534 +time/saving (s) 0.00422022 +time/training (s) 7.2839 +time/epoch (s) 10.2009 +time/total (s) 3149.84 +Epoch -696 +---------------------------------- --------------- +2022-05-10 14:03:18.875992 PDT | [0] Epoch -695 finished +---------------------------------- --------------- +epoch -695 +replay_buffer/size 999033 +trainer/num train calls 306000 +trainer/Policy Loss -19.3049 +trainer/Log Pis Mean 24.7368 +trainer/Log Pis Std 13.3472 +trainer/Log Pis Max 62.3771 +trainer/Log Pis Min -9.10685 +trainer/policy/mean Mean -0.0394485 +trainer/policy/mean Std 0.904658 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83969 +trainer/policy/normal/std Std 0.596193 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.344044 +trainer/policy/normal/log_std Mean 1.01404 +trainer/policy/normal/log_std Std 0.270046 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.06698 +eval/num steps total 303644 +eval/num paths total 306 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111692 +eval/Actions Std 0.83222 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69218 +time/logging (s) 0.00426823 +time/sampling batch (s) 0.278642 +time/saving (s) 0.00428755 +time/training (s) 7.54682 +time/epoch (s) 10.5262 +time/total (s) 3160.37 +Epoch -695 +---------------------------------- --------------- +2022-05-10 14:03:29.435421 PDT | [0] Epoch -694 finished +---------------------------------- --------------- +epoch -694 +replay_buffer/size 999033 +trainer/num train calls 307000 +trainer/Policy Loss -20.1416 +trainer/Log Pis Mean 24.5859 +trainer/Log Pis Std 13.2793 +trainer/Log Pis Max 62.3606 +trainer/Log Pis Min -8.43108 +trainer/policy/mean Mean -0.0401006 +trainer/policy/mean Std 0.902409 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81332 +trainer/policy/normal/std Std 0.611404 +trainer/policy/normal/std Max 6.03984 +trainer/policy/normal/std Min 0.349936 +trainer/policy/normal/log_std Mean 1.00057 +trainer/policy/normal/log_std Std 0.292822 +trainer/policy/normal/log_std Max 1.79838 +trainer/policy/normal/log_std Min -1.05001 +eval/num steps total 304644 +eval/num paths total 307 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0218832 +eval/Actions Std 0.904179 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60807 +time/logging (s) 0.00435243 +time/sampling batch (s) 0.281611 +time/saving (s) 0.00431127 +time/training (s) 7.64018 +time/epoch (s) 10.5385 +time/total (s) 3170.91 +Epoch -694 +---------------------------------- --------------- +2022-05-10 14:03:39.597056 PDT | [0] Epoch -693 finished +---------------------------------- --------------- +epoch -693 +replay_buffer/size 999033 +trainer/num train calls 308000 +trainer/Policy Loss -18.4644 +trainer/Log Pis Mean 23.1119 +trainer/Log Pis Std 12.9179 +trainer/Log Pis Max 66.92 +trainer/Log Pis Min -8.06668 +trainer/policy/mean Mean -0.0260626 +trainer/policy/mean Std 0.902514 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86468 +trainer/policy/normal/std Std 0.635203 +trainer/policy/normal/std Max 6.08635 +trainer/policy/normal/std Min 0.339759 +trainer/policy/normal/log_std Mean 1.017 +trainer/policy/normal/log_std Std 0.300595 +trainer/policy/normal/log_std Max 1.80605 +trainer/policy/normal/log_std Min -1.07952 +eval/num steps total 305644 +eval/num paths total 308 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0396587 +eval/Actions Std 0.909494 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33367 +time/logging (s) 0.0039953 +time/sampling batch (s) 0.535484 +time/saving (s) 0.00385593 +time/training (s) 7.2627 +time/epoch (s) 10.1397 +time/total (s) 3181.06 +Epoch -693 +---------------------------------- --------------- +2022-05-10 14:03:49.684702 PDT | [0] Epoch -692 finished +---------------------------------- --------------- +epoch -692 +replay_buffer/size 999033 +trainer/num train calls 309000 +trainer/Policy Loss -19.2372 +trainer/Log Pis Mean 25.8809 +trainer/Log Pis Std 13.0973 +trainer/Log Pis Max 65.9043 +trainer/Log Pis Min -5.4225 +trainer/policy/mean Mean -0.0333004 +trainer/policy/mean Std 0.906644 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84014 +trainer/policy/normal/std Std 0.610926 +trainer/policy/normal/std Max 4.96148 +trainer/policy/normal/std Min 0.293157 +trainer/policy/normal/log_std Mean 1.01144 +trainer/policy/normal/log_std Std 0.284244 +trainer/policy/normal/log_std Max 1.6017 +trainer/policy/normal/log_std Min -1.22705 +eval/num steps total 306644 +eval/num paths total 309 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0829291 +eval/Actions Std 0.912324 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56291 +time/logging (s) 0.00385574 +time/sampling batch (s) 0.280164 +time/saving (s) 0.00381243 +time/training (s) 7.21627 +time/epoch (s) 10.067 +time/total (s) 3191.13 +Epoch -692 +---------------------------------- --------------- +2022-05-10 14:04:00.050942 PDT | [0] Epoch -691 finished +---------------------------------- --------------- +epoch -691 +replay_buffer/size 999033 +trainer/num train calls 310000 +trainer/Policy Loss -19.557 +trainer/Log Pis Mean 24.441 +trainer/Log Pis Std 12.8976 +trainer/Log Pis Max 66.0096 +trainer/Log Pis Min -7.20292 +trainer/policy/mean Mean -0.020733 +trainer/policy/mean Std 0.903335 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85022 +trainer/policy/normal/std Std 0.612902 +trainer/policy/normal/std Max 5.38936 +trainer/policy/normal/std Min 0.406896 +trainer/policy/normal/log_std Mean 1.01552 +trainer/policy/normal/log_std Std 0.28024 +trainer/policy/normal/log_std Max 1.68443 +trainer/policy/normal/log_std Min -0.899198 +eval/num steps total 307644 +eval/num paths total 310 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.169303 +eval/Actions Std 0.873452 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64829 +time/logging (s) 0.00378941 +time/sampling batch (s) 0.281946 +time/saving (s) 0.00346844 +time/training (s) 7.40819 +time/epoch (s) 10.3457 +time/total (s) 3201.48 +Epoch -691 +---------------------------------- --------------- +2022-05-10 14:04:10.510065 PDT | [0] Epoch -690 finished +---------------------------------- --------------- +epoch -690 +replay_buffer/size 999033 +trainer/num train calls 311000 +trainer/Policy Loss -18.5553 +trainer/Log Pis Mean 25.5387 +trainer/Log Pis Std 13.5389 +trainer/Log Pis Max 65.8349 +trainer/Log Pis Min -7.63409 +trainer/policy/mean Mean -0.0276983 +trainer/policy/mean Std 0.905431 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80343 +trainer/policy/normal/std Std 0.614692 +trainer/policy/normal/std Max 5.61444 +trainer/policy/normal/std Min 0.344054 +trainer/policy/normal/log_std Mean 0.995047 +trainer/policy/normal/log_std Std 0.304133 +trainer/policy/normal/log_std Max 1.72534 +trainer/policy/normal/log_std Min -1.06696 +eval/num steps total 308644 +eval/num paths total 311 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0746414 +eval/Actions Std 0.913815 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60255 +time/logging (s) 0.00425254 +time/sampling batch (s) 0.283783 +time/saving (s) 0.0041695 +time/training (s) 7.54427 +time/epoch (s) 10.439 +time/total (s) 3211.92 +Epoch -690 +---------------------------------- --------------- +2022-05-10 14:04:21.078583 PDT | [0] Epoch -689 finished +---------------------------------- --------------- +epoch -689 +replay_buffer/size 999033 +trainer/num train calls 312000 +trainer/Policy Loss -19.6212 +trainer/Log Pis Mean 24.6252 +trainer/Log Pis Std 12.7761 +trainer/Log Pis Max 74.2996 +trainer/Log Pis Min -8.57243 +trainer/policy/mean Mean -0.0267299 +trainer/policy/mean Std 0.901932 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80585 +trainer/policy/normal/std Std 0.635613 +trainer/policy/normal/std Max 5.4268 +trainer/policy/normal/std Min 0.370582 +trainer/policy/normal/log_std Mean 0.995238 +trainer/policy/normal/log_std Std 0.301851 +trainer/policy/normal/log_std Max 1.69135 +trainer/policy/normal/log_std Min -0.992681 +eval/num steps total 309644 +eval/num paths total 312 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0564811 +eval/Actions Std 0.896645 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57786 +time/logging (s) 0.00368237 +time/sampling batch (s) 0.535305 +time/saving (s) 0.00355513 +time/training (s) 7.4262 +time/epoch (s) 10.5466 +time/total (s) 3222.47 +Epoch -689 +---------------------------------- --------------- +2022-05-10 14:04:30.480901 PDT | [0] Epoch -688 finished +---------------------------------- --------------- +epoch -688 +replay_buffer/size 999033 +trainer/num train calls 313000 +trainer/Policy Loss -19.6551 +trainer/Log Pis Mean 24.6415 +trainer/Log Pis Std 13.1967 +trainer/Log Pis Max 68.2907 +trainer/Log Pis Min -6.61752 +trainer/policy/mean Mean -0.0326023 +trainer/policy/mean Std 0.90556 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81283 +trainer/policy/normal/std Std 0.615529 +trainer/policy/normal/std Max 5.96858 +trainer/policy/normal/std Min 0.416681 +trainer/policy/normal/log_std Mean 1.00178 +trainer/policy/normal/log_std Std 0.281461 +trainer/policy/normal/log_std Max 1.78651 +trainer/policy/normal/log_std Min -0.875435 +eval/num steps total 310644 +eval/num paths total 313 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179713 +eval/Actions Std 0.852991 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55936 +time/logging (s) 0.00383442 +time/sampling batch (s) 0.283344 +time/saving (s) 0.00358045 +time/training (s) 6.53195 +time/epoch (s) 9.38206 +time/total (s) 3231.85 +Epoch -688 +---------------------------------- --------------- +2022-05-10 14:04:39.917186 PDT | [0] Epoch -687 finished +---------------------------------- --------------- +epoch -687 +replay_buffer/size 999033 +trainer/num train calls 314000 +trainer/Policy Loss -19.2308 +trainer/Log Pis Mean 24.5942 +trainer/Log Pis Std 13.7186 +trainer/Log Pis Max 71.6946 +trainer/Log Pis Min -6.42284 +trainer/policy/mean Mean -0.0259241 +trainer/policy/mean Std 0.905286 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.86524 +trainer/policy/normal/std Std 0.64362 +trainer/policy/normal/std Max 5.39409 +trainer/policy/normal/std Min 0.311548 +trainer/policy/normal/log_std Mean 1.01534 +trainer/policy/normal/log_std Std 0.310781 +trainer/policy/normal/log_std Max 1.6853 +trainer/policy/normal/log_std Min -1.1662 +eval/num steps total 311644 +eval/num paths total 314 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.137711 +eval/Actions Std 0.908728 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62204 +time/logging (s) 0.00375152 +time/sampling batch (s) 0.28222 +time/saving (s) 0.00345868 +time/training (s) 6.50438 +time/epoch (s) 9.41585 +time/total (s) 3241.27 +Epoch -687 +---------------------------------- --------------- +2022-05-10 14:04:50.826527 PDT | [0] Epoch -686 finished +---------------------------------- --------------- +epoch -686 +replay_buffer/size 999033 +trainer/num train calls 315000 +trainer/Policy Loss -18.6831 +trainer/Log Pis Mean 24.3553 +trainer/Log Pis Std 13.8007 +trainer/Log Pis Max 74.0427 +trainer/Log Pis Min -10.788 +trainer/policy/mean Mean -0.0294894 +trainer/policy/mean Std 0.908184 +trainer/policy/mean Max 0.999971 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81956 +trainer/policy/normal/std Std 0.613113 +trainer/policy/normal/std Max 5.98192 +trainer/policy/normal/std Min 0.340565 +trainer/policy/normal/log_std Mean 1.00263 +trainer/policy/normal/log_std Std 0.29313 +trainer/policy/normal/log_std Max 1.78874 +trainer/policy/normal/log_std Min -1.07715 +eval/num steps total 312644 +eval/num paths total 315 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.155075 +eval/Actions Std 0.945844 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40481 +time/logging (s) 0.00380072 +time/sampling batch (s) 0.294313 +time/saving (s) 0.00345478 +time/training (s) 8.18205 +time/epoch (s) 10.8884 +time/total (s) 3252.16 +Epoch -686 +---------------------------------- --------------- +2022-05-10 14:05:00.727107 PDT | [0] Epoch -685 finished +---------------------------------- --------------- +epoch -685 +replay_buffer/size 999033 +trainer/num train calls 316000 +trainer/Policy Loss -20.1952 +trainer/Log Pis Mean 25.1916 +trainer/Log Pis Std 12.9938 +trainer/Log Pis Max 70.2284 +trainer/Log Pis Min -7.38662 +trainer/policy/mean Mean -0.03912 +trainer/policy/mean Std 0.909803 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.90665 +trainer/policy/normal/std Std 0.603957 +trainer/policy/normal/std Max 5.97758 +trainer/policy/normal/std Min 0.412992 +trainer/policy/normal/log_std Mean 1.03676 +trainer/policy/normal/log_std Std 0.275114 +trainer/policy/normal/log_std Max 1.78802 +trainer/policy/normal/log_std Min -0.884327 +eval/num steps total 313644 +eval/num paths total 316 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0960173 +eval/Actions Std 0.899788 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67558 +time/logging (s) 0.00416988 +time/sampling batch (s) 0.284124 +time/saving (s) 0.00410299 +time/training (s) 6.91232 +time/epoch (s) 9.8803 +time/total (s) 3262.05 +Epoch -685 +---------------------------------- --------------- +2022-05-10 14:05:10.070573 PDT | [0] Epoch -684 finished +---------------------------------- --------------- +epoch -684 +replay_buffer/size 999033 +trainer/num train calls 317000 +trainer/Policy Loss -19.2224 +trainer/Log Pis Mean 23.4873 +trainer/Log Pis Std 13.4689 +trainer/Log Pis Max 62.2284 +trainer/Log Pis Min -8.47786 +trainer/policy/mean Mean -0.0594208 +trainer/policy/mean Std 0.904795 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84226 +trainer/policy/normal/std Std 0.652945 +trainer/policy/normal/std Max 5.26201 +trainer/policy/normal/std Min 0.315709 +trainer/policy/normal/log_std Mean 1.00504 +trainer/policy/normal/log_std Std 0.319783 +trainer/policy/normal/log_std Max 1.66051 +trainer/policy/normal/log_std Min -1.15294 +eval/num steps total 314644 +eval/num paths total 317 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0184789 +eval/Actions Std 0.903901 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.27409 +time/logging (s) 0.00382367 +time/sampling batch (s) 0.284532 +time/saving (s) 0.00365133 +time/training (s) 6.75602 +time/epoch (s) 9.32212 +time/total (s) 3271.37 +Epoch -684 +---------------------------------- --------------- +2022-05-10 14:05:19.766536 PDT | [0] Epoch -683 finished +---------------------------------- --------------- +epoch -683 +replay_buffer/size 999033 +trainer/num train calls 318000 +trainer/Policy Loss -19.5724 +trainer/Log Pis Mean 24.7718 +trainer/Log Pis Std 13.6196 +trainer/Log Pis Max 71.4401 +trainer/Log Pis Min -7.38423 +trainer/policy/mean Mean -0.0438646 +trainer/policy/mean Std 0.906988 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78846 +trainer/policy/normal/std Std 0.647391 +trainer/policy/normal/std Max 6.7356 +trainer/policy/normal/std Min 0.321715 +trainer/policy/normal/log_std Mean 0.986693 +trainer/policy/normal/log_std Std 0.315262 +trainer/policy/normal/log_std Max 1.90741 +trainer/policy/normal/log_std Min -1.13409 +eval/num steps total 315644 +eval/num paths total 318 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.067672 +eval/Actions Std 0.905863 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52852 +time/logging (s) 0.00373126 +time/sampling batch (s) 0.282865 +time/saving (s) 0.00357297 +time/training (s) 6.85657 +time/epoch (s) 9.67526 +time/total (s) 3281.05 +Epoch -683 +---------------------------------- --------------- +2022-05-10 14:05:30.411175 PDT | [0] Epoch -682 finished +---------------------------------- --------------- +epoch -682 +replay_buffer/size 999033 +trainer/num train calls 319000 +trainer/Policy Loss -20.258 +trainer/Log Pis Mean 24.8153 +trainer/Log Pis Std 13.4451 +trainer/Log Pis Max 71.2813 +trainer/Log Pis Min -8.76388 +trainer/policy/mean Mean -0.0335365 +trainer/policy/mean Std 0.906333 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87759 +trainer/policy/normal/std Std 0.638934 +trainer/policy/normal/std Max 5.89472 +trainer/policy/normal/std Min 0.345048 +trainer/policy/normal/log_std Mean 1.02204 +trainer/policy/normal/log_std Std 0.296642 +trainer/policy/normal/log_std Max 1.77406 +trainer/policy/normal/log_std Min -1.06407 +eval/num steps total 316644 +eval/num paths total 319 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0917464 +eval/Actions Std 0.896205 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60472 +time/logging (s) 0.00366611 +time/sampling batch (s) 0.280656 +time/saving (s) 0.00341456 +time/training (s) 7.73189 +time/epoch (s) 10.6243 +time/total (s) 3291.68 +Epoch -682 +---------------------------------- --------------- +2022-05-10 14:05:40.898468 PDT | [0] Epoch -681 finished +---------------------------------- --------------- +epoch -681 +replay_buffer/size 999033 +trainer/num train calls 320000 +trainer/Policy Loss -19.1602 +trainer/Log Pis Mean 23.5964 +trainer/Log Pis Std 12.8966 +trainer/Log Pis Max 73.7674 +trainer/Log Pis Min -8.70304 +trainer/policy/mean Mean -0.0487812 +trainer/policy/mean Std 0.900998 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79897 +trainer/policy/normal/std Std 0.636732 +trainer/policy/normal/std Max 6.6501 +trainer/policy/normal/std Min 0.378775 +trainer/policy/normal/log_std Mean 0.992764 +trainer/policy/normal/log_std Std 0.302188 +trainer/policy/normal/log_std Max 1.89463 +trainer/policy/normal/log_std Min -0.970812 +eval/num steps total 317644 +eval/num paths total 320 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.134682 +eval/Actions Std 0.902916 +eval/Actions Max 0.999981 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66787 +time/logging (s) 0.00368036 +time/sampling batch (s) 0.280037 +time/saving (s) 0.00349813 +time/training (s) 7.51196 +time/epoch (s) 10.4671 +time/total (s) 3302.15 +Epoch -681 +---------------------------------- --------------- +2022-05-10 14:05:52.108402 PDT | [0] Epoch -680 finished +---------------------------------- --------------- +epoch -680 +replay_buffer/size 999033 +trainer/num train calls 321000 +trainer/Policy Loss -19.5535 +trainer/Log Pis Mean 24.4776 +trainer/Log Pis Std 13.5075 +trainer/Log Pis Max 63.034 +trainer/Log Pis Min -10.5551 +trainer/policy/mean Mean -0.0386105 +trainer/policy/mean Std 0.903613 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.84543 +trainer/policy/normal/std Std 0.611058 +trainer/policy/normal/std Max 5.49383 +trainer/policy/normal/std Min 0.332321 +trainer/policy/normal/log_std Mean 1.01318 +trainer/policy/normal/log_std Std 0.286005 +trainer/policy/normal/log_std Max 1.70363 +trainer/policy/normal/log_std Min -1.10165 +eval/num steps total 318644 +eval/num paths total 321 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110326 +eval/Actions Std 0.908588 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6238 +time/logging (s) 0.00374922 +time/sampling batch (s) 0.533648 +time/saving (s) 0.00348879 +time/training (s) 8.02504 +time/epoch (s) 11.1897 +time/total (s) 3313.34 +Epoch -680 +---------------------------------- --------------- +2022-05-10 14:06:01.703375 PDT | [0] Epoch -679 finished +---------------------------------- --------------- +epoch -679 +replay_buffer/size 999033 +trainer/num train calls 322000 +trainer/Policy Loss -19.2699 +trainer/Log Pis Mean 24.0121 +trainer/Log Pis Std 12.8965 +trainer/Log Pis Max 61.3459 +trainer/Log Pis Min -8.35471 +trainer/policy/mean Mean -0.0229959 +trainer/policy/mean Std 0.903299 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81657 +trainer/policy/normal/std Std 0.611665 +trainer/policy/normal/std Max 5.79294 +trainer/policy/normal/std Min 0.330346 +trainer/policy/normal/log_std Mean 1.00242 +trainer/policy/normal/log_std Std 0.287659 +trainer/policy/normal/log_std Max 1.75664 +trainer/policy/normal/log_std Min -1.10761 +eval/num steps total 319644 +eval/num paths total 322 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00129881 +eval/Actions Std 0.907427 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48916 +time/logging (s) 0.00403685 +time/sampling batch (s) 0.280845 +time/saving (s) 0.00368569 +time/training (s) 6.7972 +time/epoch (s) 9.57493 +time/total (s) 3322.92 +Epoch -679 +---------------------------------- --------------- +2022-05-10 14:06:11.499731 PDT | [0] Epoch -678 finished +---------------------------------- --------------- +epoch -678 +replay_buffer/size 999033 +trainer/num train calls 323000 +trainer/Policy Loss -20.3273 +trainer/Log Pis Mean 24.255 +trainer/Log Pis Std 13.6336 +trainer/Log Pis Max 65.181 +trainer/Log Pis Min -7.1741 +trainer/policy/mean Mean -0.0273149 +trainer/policy/mean Std 0.908155 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85694 +trainer/policy/normal/std Std 0.633965 +trainer/policy/normal/std Max 5.36679 +trainer/policy/normal/std Min 0.307308 +trainer/policy/normal/log_std Mean 1.01464 +trainer/policy/normal/log_std Std 0.297708 +trainer/policy/normal/log_std Max 1.68023 +trainer/policy/normal/log_std Min -1.1799 +eval/num steps total 320227 +eval/num paths total 323 +eval/path length Mean 583 +eval/path length Std 0 +eval/path length Max 583 +eval/path length Min 583 +eval/Rewards Mean 0.00171527 +eval/Rewards Std 0.0413802 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.019503 +eval/Actions Std 0.902239 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.6713 +time/logging (s) 0.00302073 +time/sampling batch (s) 0.282996 +time/saving (s) 0.00405509 +time/training (s) 6.81319 +time/epoch (s) 9.77456 +time/total (s) 3332.69 +Epoch -678 +---------------------------------- --------------- +2022-05-10 14:06:22.309395 PDT | [0] Epoch -677 finished +---------------------------------- --------------- +epoch -677 +replay_buffer/size 999033 +trainer/num train calls 324000 +trainer/Policy Loss -19.6877 +trainer/Log Pis Mean 23.9206 +trainer/Log Pis Std 13.2969 +trainer/Log Pis Max 69.9638 +trainer/Log Pis Min -7.04459 +trainer/policy/mean Mean -0.0272108 +trainer/policy/mean Std 0.906048 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.87512 +trainer/policy/normal/std Std 0.61906 +trainer/policy/normal/std Max 6.63017 +trainer/policy/normal/std Min 0.429341 +trainer/policy/normal/log_std Mean 1.02365 +trainer/policy/normal/log_std Std 0.283986 +trainer/policy/normal/log_std Max 1.89163 +trainer/policy/normal/log_std Min -0.845503 +eval/num steps total 321227 +eval/num paths total 324 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112803 +eval/Actions Std 0.89424 +eval/Actions Max 0.999981 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73518 +time/logging (s) 0.0037762 +time/sampling batch (s) 0.530897 +time/saving (s) 0.00348836 +time/training (s) 7.51642 +time/epoch (s) 10.7898 +time/total (s) 3343.49 +Epoch -677 +---------------------------------- --------------- +2022-05-10 14:06:31.922008 PDT | [0] Epoch -676 finished +---------------------------------- --------------- +epoch -676 +replay_buffer/size 999033 +trainer/num train calls 325000 +trainer/Policy Loss -19.6504 +trainer/Log Pis Mean 24.7793 +trainer/Log Pis Std 12.8857 +trainer/Log Pis Max 61.1183 +trainer/Log Pis Min -8.1194 +trainer/policy/mean Mean -0.0469272 +trainer/policy/mean Std 0.909011 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82647 +trainer/policy/normal/std Std 0.594891 +trainer/policy/normal/std Max 5.11242 +trainer/policy/normal/std Min 0.415099 +trainer/policy/normal/log_std Mean 1.0082 +trainer/policy/normal/log_std Std 0.276792 +trainer/policy/normal/log_std Max 1.63167 +trainer/policy/normal/log_std Min -0.879238 +eval/num steps total 322227 +eval/num paths total 325 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0068917 +eval/Actions Std 0.867105 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72126 +time/logging (s) 0.00370265 +time/sampling batch (s) 0.281523 +time/saving (s) 0.00350064 +time/training (s) 6.58198 +time/epoch (s) 9.59197 +time/total (s) 3353.08 +Epoch -676 +---------------------------------- --------------- +2022-05-10 14:06:41.735732 PDT | [0] Epoch -675 finished +---------------------------------- --------------- +epoch -675 +replay_buffer/size 999033 +trainer/num train calls 326000 +trainer/Policy Loss -19.1625 +trainer/Log Pis Mean 23.7214 +trainer/Log Pis Std 13.1213 +trainer/Log Pis Max 61.2103 +trainer/Log Pis Min -10.9648 +trainer/policy/mean Mean -0.0268904 +trainer/policy/mean Std 0.90708 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8385 +trainer/policy/normal/std Std 0.608158 +trainer/policy/normal/std Max 5.4416 +trainer/policy/normal/std Min 0.335006 +trainer/policy/normal/log_std Mean 1.00992 +trainer/policy/normal/log_std Std 0.291385 +trainer/policy/normal/log_std Max 1.69407 +trainer/policy/normal/log_std Min -1.09361 +eval/num steps total 323227 +eval/num paths total 326 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122981 +eval/Actions Std 0.870818 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54461 +time/logging (s) 0.00369459 +time/sampling batch (s) 0.283978 +time/saving (s) 0.00343254 +time/training (s) 6.9575 +time/epoch (s) 9.79321 +time/total (s) 3362.88 +Epoch -675 +---------------------------------- --------------- +2022-05-10 14:06:52.698301 PDT | [0] Epoch -674 finished +---------------------------------- --------------- +epoch -674 +replay_buffer/size 999033 +trainer/num train calls 327000 +trainer/Policy Loss -19.7392 +trainer/Log Pis Mean 24.5927 +trainer/Log Pis Std 14.0419 +trainer/Log Pis Max 71.7217 +trainer/Log Pis Min -5.07837 +trainer/policy/mean Mean -0.048888 +trainer/policy/mean Std 0.90438 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.87585 +trainer/policy/normal/std Std 0.615707 +trainer/policy/normal/std Max 5.0254 +trainer/policy/normal/std Min 0.360252 +trainer/policy/normal/log_std Mean 1.02416 +trainer/policy/normal/log_std Std 0.283764 +trainer/policy/normal/log_std Max 1.6145 +trainer/policy/normal/log_std Min -1.02095 +eval/num steps total 324227 +eval/num paths total 327 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.191087 +eval/Actions Std 0.920686 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67803 +time/logging (s) 0.00368591 +time/sampling batch (s) 0.782753 +time/saving (s) 0.00345676 +time/training (s) 7.4741 +time/epoch (s) 10.942 +time/total (s) 3373.82 +Epoch -674 +---------------------------------- --------------- +2022-05-10 14:07:03.977928 PDT | [0] Epoch -673 finished +---------------------------------- --------------- +epoch -673 +replay_buffer/size 999033 +trainer/num train calls 328000 +trainer/Policy Loss -19.1216 +trainer/Log Pis Mean 24.2787 +trainer/Log Pis Std 13.2708 +trainer/Log Pis Max 65.4105 +trainer/Log Pis Min -8.00175 +trainer/policy/mean Mean -0.0545532 +trainer/policy/mean Std 0.902728 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83077 +trainer/policy/normal/std Std 0.619657 +trainer/policy/normal/std Max 5.46994 +trainer/policy/normal/std Min 0.38419 +trainer/policy/normal/log_std Mean 1.00627 +trainer/policy/normal/log_std Std 0.294595 +trainer/policy/normal/log_std Max 1.69927 +trainer/policy/normal/log_std Min -0.956618 +eval/num steps total 325227 +eval/num paths total 328 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0874815 +eval/Actions Std 0.915682 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61782 +time/logging (s) 0.00418843 +time/sampling batch (s) 0.53331 +time/saving (s) 0.00418137 +time/training (s) 8.09997 +time/epoch (s) 11.2595 +time/total (s) 3385.09 +Epoch -673 +---------------------------------- --------------- +2022-05-10 14:07:13.843716 PDT | [0] Epoch -672 finished +---------------------------------- --------------- +epoch -672 +replay_buffer/size 999033 +trainer/num train calls 329000 +trainer/Policy Loss -19.1226 +trainer/Log Pis Mean 24.9172 +trainer/Log Pis Std 13.1217 +trainer/Log Pis Max 62.6617 +trainer/Log Pis Min -5.17913 +trainer/policy/mean Mean -0.0472778 +trainer/policy/mean Std 0.910374 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.7714 +trainer/policy/normal/std Std 0.617481 +trainer/policy/normal/std Max 5.27052 +trainer/policy/normal/std Min 0.368578 +trainer/policy/normal/log_std Mean 0.983465 +trainer/policy/normal/log_std Std 0.301813 +trainer/policy/normal/log_std Max 1.66213 +trainer/policy/normal/log_std Min -0.998102 +eval/num steps total 326227 +eval/num paths total 329 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.431164 +eval/Actions Std 0.888029 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72161 +time/logging (s) 0.00381468 +time/sampling batch (s) 0.283149 +time/saving (s) 0.00361031 +time/training (s) 6.83198 +time/epoch (s) 9.84417 +time/total (s) 3394.93 +Epoch -672 +---------------------------------- --------------- +2022-05-10 14:07:24.758013 PDT | [0] Epoch -671 finished +---------------------------------- --------------- +epoch -671 +replay_buffer/size 999033 +trainer/num train calls 330000 +trainer/Policy Loss -20.6748 +trainer/Log Pis Mean 24.4529 +trainer/Log Pis Std 13.0668 +trainer/Log Pis Max 67.8375 +trainer/Log Pis Min -11.112 +trainer/policy/mean Mean -0.0205886 +trainer/policy/mean Std 0.907536 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.83773 +trainer/policy/normal/std Std 0.624541 +trainer/policy/normal/std Max 5.33416 +trainer/policy/normal/std Min 0.335285 +trainer/policy/normal/log_std Mean 1.00815 +trainer/policy/normal/log_std Std 0.297786 +trainer/policy/normal/log_std Max 1.67413 +trainer/policy/normal/log_std Min -1.09278 +eval/num steps total 327227 +eval/num paths total 330 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.195376 +eval/Actions Std 0.903952 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62308 +time/logging (s) 0.00368179 +time/sampling batch (s) 0.283306 +time/saving (s) 0.0034556 +time/training (s) 7.97979 +time/epoch (s) 10.8933 +time/total (s) 3405.83 +Epoch -671 +---------------------------------- --------------- +2022-05-10 14:07:35.758017 PDT | [0] Epoch -670 finished +---------------------------------- --------------- +epoch -670 +replay_buffer/size 999033 +trainer/num train calls 331000 +trainer/Policy Loss -19.2821 +trainer/Log Pis Mean 23.8779 +trainer/Log Pis Std 12.6721 +trainer/Log Pis Max 62.4332 +trainer/Log Pis Min -10.22 +trainer/policy/mean Mean -0.0214257 +trainer/policy/mean Std 0.903448 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.89543 +trainer/policy/normal/std Std 0.618437 +trainer/policy/normal/std Max 5.40604 +trainer/policy/normal/std Min 0.328691 +trainer/policy/normal/log_std Mean 1.03129 +trainer/policy/normal/log_std Std 0.282212 +trainer/policy/normal/log_std Max 1.68752 +trainer/policy/normal/log_std Min -1.11264 +eval/num steps total 328227 +eval/num paths total 331 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.200518 +eval/Actions Std 0.956672 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51643 +time/logging (s) 0.00398566 +time/sampling batch (s) 0.784286 +time/saving (s) 0.00355087 +time/training (s) 7.67123 +time/epoch (s) 10.9795 +time/total (s) 3416.81 +Epoch -670 +---------------------------------- --------------- +2022-05-10 14:07:45.238633 PDT | [0] Epoch -669 finished +---------------------------------- --------------- +epoch -669 +replay_buffer/size 999033 +trainer/num train calls 332000 +trainer/Policy Loss -18.929 +trainer/Log Pis Mean 23.4798 +trainer/Log Pis Std 13.6558 +trainer/Log Pis Max 63.7955 +trainer/Log Pis Min -9.29042 +trainer/policy/mean Mean -0.00337705 +trainer/policy/mean Std 0.904779 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84744 +trainer/policy/normal/std Std 0.62419 +trainer/policy/normal/std Max 5.07068 +trainer/policy/normal/std Min 0.330706 +trainer/policy/normal/log_std Mean 1.01254 +trainer/policy/normal/log_std Std 0.291404 +trainer/policy/normal/log_std Max 1.62347 +trainer/policy/normal/log_std Min -1.10653 +eval/num steps total 329227 +eval/num paths total 332 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0026153 +eval/Actions Std 0.968149 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57693 +time/logging (s) 0.00375666 +time/sampling batch (s) 0.287977 +time/saving (s) 0.00349579 +time/training (s) 6.58723 +time/epoch (s) 9.45939 +time/total (s) 3426.27 +Epoch -669 +---------------------------------- --------------- +2022-05-10 14:07:54.972883 PDT | [0] Epoch -668 finished +---------------------------------- --------------- +epoch -668 +replay_buffer/size 999033 +trainer/num train calls 333000 +trainer/Policy Loss -18.9799 +trainer/Log Pis Mean 24.8477 +trainer/Log Pis Std 14.0065 +trainer/Log Pis Max 78.2726 +trainer/Log Pis Min -9.9963 +trainer/policy/mean Mean -0.0327096 +trainer/policy/mean Std 0.907897 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.77446 +trainer/policy/normal/std Std 0.636773 +trainer/policy/normal/std Max 6.89532 +trainer/policy/normal/std Min 0.351566 +trainer/policy/normal/log_std Mean 0.981533 +trainer/policy/normal/log_std Std 0.316188 +trainer/policy/normal/log_std Max 1.93084 +trainer/policy/normal/log_std Min -1.04536 +eval/num steps total 330227 +eval/num paths total 333 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.220202 +eval/Actions Std 0.89315 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70258 +time/logging (s) 0.00376142 +time/sampling batch (s) 0.279356 +time/saving (s) 0.00361361 +time/training (s) 6.72457 +time/epoch (s) 9.71388 +time/total (s) 3435.99 +Epoch -668 +---------------------------------- --------------- +2022-05-10 14:08:05.374624 PDT | [0] Epoch -667 finished +---------------------------------- --------------- +epoch -667 +replay_buffer/size 999033 +trainer/num train calls 334000 +trainer/Policy Loss -20.0269 +trainer/Log Pis Mean 23.7206 +trainer/Log Pis Std 13.3396 +trainer/Log Pis Max 65.1952 +trainer/Log Pis Min -7.22409 +trainer/policy/mean Mean -0.0404445 +trainer/policy/mean Std 0.90737 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78088 +trainer/policy/normal/std Std 0.59763 +trainer/policy/normal/std Max 5.40929 +trainer/policy/normal/std Min 0.345565 +trainer/policy/normal/log_std Mean 0.989975 +trainer/policy/normal/log_std Std 0.286869 +trainer/policy/normal/log_std Max 1.68812 +trainer/policy/normal/log_std Min -1.06257 +eval/num steps total 331227 +eval/num paths total 334 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0110762 +eval/Actions Std 0.904382 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55565 +time/logging (s) 0.00396129 +time/sampling batch (s) 0.52916 +time/saving (s) 0.00398627 +time/training (s) 7.28884 +time/epoch (s) 10.3816 +time/total (s) 3446.37 +Epoch -667 +---------------------------------- --------------- +2022-05-10 14:08:15.361886 PDT | [0] Epoch -666 finished +---------------------------------- --------------- +epoch -666 +replay_buffer/size 999033 +trainer/num train calls 335000 +trainer/Policy Loss -20.6406 +trainer/Log Pis Mean 25.6042 +trainer/Log Pis Std 13.5154 +trainer/Log Pis Max 67.3554 +trainer/Log Pis Min -5.86397 +trainer/policy/mean Mean -0.0235997 +trainer/policy/mean Std 0.904741 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.90002 +trainer/policy/normal/std Std 0.686078 +trainer/policy/normal/std Max 5.37165 +trainer/policy/normal/std Min 0.404928 +trainer/policy/normal/log_std Mean 1.02573 +trainer/policy/normal/log_std Std 0.311248 +trainer/policy/normal/log_std Max 1.68113 +trainer/policy/normal/log_std Min -0.904045 +eval/num steps total 332227 +eval/num paths total 335 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.298512 +eval/Actions Std 0.897319 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71072 +time/logging (s) 0.00371632 +time/sampling batch (s) 0.280902 +time/saving (s) 0.00347729 +time/training (s) 6.9674 +time/epoch (s) 9.96622 +time/total (s) 3456.34 +Epoch -666 +---------------------------------- --------------- +2022-05-10 14:08:25.060523 PDT | [0] Epoch -665 finished +---------------------------------- --------------- +epoch -665 +replay_buffer/size 999033 +trainer/num train calls 336000 +trainer/Policy Loss -20.3437 +trainer/Log Pis Mean 25.0026 +trainer/Log Pis Std 13.4586 +trainer/Log Pis Max 73.4845 +trainer/Log Pis Min -7.60158 +trainer/policy/mean Mean -0.0564 +trainer/policy/mean Std 0.906753 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82303 +trainer/policy/normal/std Std 0.645702 +trainer/policy/normal/std Max 5.90607 +trainer/policy/normal/std Min 0.369777 +trainer/policy/normal/log_std Mean 1.0003 +trainer/policy/normal/log_std Std 0.308456 +trainer/policy/normal/log_std Max 1.77598 +trainer/policy/normal/log_std Min -0.994855 +eval/num steps total 333227 +eval/num paths total 336 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0472293 +eval/Actions Std 0.895855 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50927 +time/logging (s) 0.00366357 +time/sampling batch (s) 0.533026 +time/saving (s) 0.00341481 +time/training (s) 6.62873 +time/epoch (s) 9.6781 +time/total (s) 3466.03 +Epoch -665 +---------------------------------- --------------- +2022-05-10 14:08:34.587971 PDT | [0] Epoch -664 finished +---------------------------------- --------------- +epoch -664 +replay_buffer/size 999033 +trainer/num train calls 337000 +trainer/Policy Loss -20.1532 +trainer/Log Pis Mean 24.9187 +trainer/Log Pis Std 14.1294 +trainer/Log Pis Max 64.9002 +trainer/Log Pis Min -6.87904 +trainer/policy/mean Mean -0.0271798 +trainer/policy/mean Std 0.911829 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8707 +trainer/policy/normal/std Std 0.630768 +trainer/policy/normal/std Max 5.44085 +trainer/policy/normal/std Min 0.354829 +trainer/policy/normal/log_std Mean 1.01957 +trainer/policy/normal/log_std Std 0.2982 +trainer/policy/normal/log_std Max 1.69394 +trainer/policy/normal/log_std Min -1.03612 +eval/num steps total 334227 +eval/num paths total 337 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12347 +eval/Actions Std 0.880859 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65536 +time/logging (s) 0.00365553 +time/sampling batch (s) 0.278602 +time/saving (s) 0.00343302 +time/training (s) 6.56628 +time/epoch (s) 9.50733 +time/total (s) 3475.54 +Epoch -664 +---------------------------------- --------------- +2022-05-10 14:08:45.178767 PDT | [0] Epoch -663 finished +---------------------------------- --------------- +epoch -663 +replay_buffer/size 999033 +trainer/num train calls 338000 +trainer/Policy Loss -19.8343 +trainer/Log Pis Mean 24.8315 +trainer/Log Pis Std 12.8782 +trainer/Log Pis Max 74.3955 +trainer/Log Pis Min -12.9012 +trainer/policy/mean Mean -0.0269268 +trainer/policy/mean Std 0.905532 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.856 +trainer/policy/normal/std Std 0.638119 +trainer/policy/normal/std Max 6.14999 +trainer/policy/normal/std Min 0.301006 +trainer/policy/normal/log_std Mean 1.01309 +trainer/policy/normal/log_std Std 0.30488 +trainer/policy/normal/log_std Max 1.81645 +trainer/policy/normal/log_std Min -1.20063 +eval/num steps total 335227 +eval/num paths total 338 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0561594 +eval/Actions Std 0.906059 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50714 +time/logging (s) 0.00366778 +time/sampling batch (s) 0.286843 +time/saving (s) 0.00343665 +time/training (s) 7.76905 +time/epoch (s) 10.5701 +time/total (s) 3486.11 +Epoch -663 +---------------------------------- --------------- +2022-05-10 14:08:54.913957 PDT | [0] Epoch -662 finished +---------------------------------- --------------- +epoch -662 +replay_buffer/size 999033 +trainer/num train calls 339000 +trainer/Policy Loss -18.4186 +trainer/Log Pis Mean 24.1527 +trainer/Log Pis Std 12.6563 +trainer/Log Pis Max 64.948 +trainer/Log Pis Min -9.06277 +trainer/policy/mean Mean -0.0417142 +trainer/policy/mean Std 0.904886 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82643 +trainer/policy/normal/std Std 0.643935 +trainer/policy/normal/std Max 5.31273 +trainer/policy/normal/std Min 0.319083 +trainer/policy/normal/log_std Mean 1.00112 +trainer/policy/normal/log_std Std 0.311989 +trainer/policy/normal/log_std Max 1.67011 +trainer/policy/normal/log_std Min -1.1423 +eval/num steps total 336227 +eval/num paths total 339 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0197055 +eval/Actions Std 0.904178 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63649 +time/logging (s) 0.00417319 +time/sampling batch (s) 0.280292 +time/saving (s) 0.00416993 +time/training (s) 6.78981 +time/epoch (s) 9.71494 +time/total (s) 3495.83 +Epoch -662 +---------------------------------- --------------- +2022-05-10 14:09:05.023616 PDT | [0] Epoch -661 finished +---------------------------------- --------------- +epoch -661 +replay_buffer/size 999033 +trainer/num train calls 340000 +trainer/Policy Loss -19.9741 +trainer/Log Pis Mean 23.3264 +trainer/Log Pis Std 12.4248 +trainer/Log Pis Max 63.6685 +trainer/Log Pis Min -7.608 +trainer/policy/mean Mean -0.025563 +trainer/policy/mean Std 0.907758 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.74324 +trainer/policy/normal/std Std 0.603967 +trainer/policy/normal/std Max 5.33816 +trainer/policy/normal/std Min 0.388315 +trainer/policy/normal/log_std Mean 0.974496 +trainer/policy/normal/log_std Std 0.295042 +trainer/policy/normal/log_std Max 1.67488 +trainer/policy/normal/log_std Min -0.945939 +eval/num steps total 337227 +eval/num paths total 340 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0130931 +eval/Actions Std 0.913955 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51079 +time/logging (s) 0.0039179 +time/sampling batch (s) 0.290474 +time/saving (s) 0.00372039 +time/training (s) 7.27896 +time/epoch (s) 10.0879 +time/total (s) 3505.92 +Epoch -661 +---------------------------------- --------------- +2022-05-10 14:09:14.942748 PDT | [0] Epoch -660 finished +---------------------------------- --------------- +epoch -660 +replay_buffer/size 999033 +trainer/num train calls 341000 +trainer/Policy Loss -19.7344 +trainer/Log Pis Mean 24.2547 +trainer/Log Pis Std 12.7346 +trainer/Log Pis Max 60.6162 +trainer/Log Pis Min -5.25287 +trainer/policy/mean Mean -0.0375633 +trainer/policy/mean Std 0.903215 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85365 +trainer/policy/normal/std Std 0.627329 +trainer/policy/normal/std Max 5.70246 +trainer/policy/normal/std Min 0.392578 +trainer/policy/normal/log_std Mean 1.01322 +trainer/policy/normal/log_std Std 0.300689 +trainer/policy/normal/log_std Max 1.7409 +trainer/policy/normal/log_std Min -0.93502 +eval/num steps total 338227 +eval/num paths total 341 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0774748 +eval/Actions Std 0.915979 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77698 +time/logging (s) 0.00383901 +time/sampling batch (s) 0.292083 +time/saving (s) 0.00353311 +time/training (s) 6.82132 +time/epoch (s) 9.89775 +time/total (s) 3515.82 +Epoch -660 +---------------------------------- --------------- +2022-05-10 14:09:25.496174 PDT | [0] Epoch -659 finished +---------------------------------- --------------- +epoch -659 +replay_buffer/size 999033 +trainer/num train calls 342000 +trainer/Policy Loss -19.1464 +trainer/Log Pis Mean 24.905 +trainer/Log Pis Std 12.9184 +trainer/Log Pis Max 67.5101 +trainer/Log Pis Min -6.64517 +trainer/policy/mean Mean -0.0206647 +trainer/policy/mean Std 0.901875 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.79793 +trainer/policy/normal/std Std 0.621301 +trainer/policy/normal/std Max 4.72939 +trainer/policy/normal/std Min 0.334995 +trainer/policy/normal/log_std Mean 0.992102 +trainer/policy/normal/log_std Std 0.308556 +trainer/policy/normal/log_std Max 1.5538 +trainer/policy/normal/log_std Min -1.09364 +eval/num steps total 339227 +eval/num paths total 342 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0103039 +eval/Actions Std 0.940557 +eval/Actions Max 0.999998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84184 +time/logging (s) 0.00382379 +time/sampling batch (s) 0.290503 +time/saving (s) 0.00355399 +time/training (s) 7.39239 +time/epoch (s) 10.5321 +time/total (s) 3526.35 +Epoch -659 +---------------------------------- --------------- +2022-05-10 14:09:34.922267 PDT | [0] Epoch -658 finished +---------------------------------- --------------- +epoch -658 +replay_buffer/size 999033 +trainer/num train calls 343000 +trainer/Policy Loss -19.1312 +trainer/Log Pis Mean 24.8537 +trainer/Log Pis Std 13.6781 +trainer/Log Pis Max 82.5305 +trainer/Log Pis Min -7.23871 +trainer/policy/mean Mean -0.0256864 +trainer/policy/mean Std 0.909019 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83154 +trainer/policy/normal/std Std 0.599939 +trainer/policy/normal/std Max 6.46177 +trainer/policy/normal/std Min 0.345354 +trainer/policy/normal/log_std Mean 1.01094 +trainer/policy/normal/log_std Std 0.268883 +trainer/policy/normal/log_std Max 1.8659 +trainer/policy/normal/log_std Min -1.06318 +eval/num steps total 340227 +eval/num paths total 343 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.113023 +eval/Actions Std 0.891182 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75016 +time/logging (s) 0.00381069 +time/sampling batch (s) 0.291105 +time/saving (s) 0.00356408 +time/training (s) 6.35619 +time/epoch (s) 9.40483 +time/total (s) 3535.76 +Epoch -658 +---------------------------------- --------------- +2022-05-10 14:09:44.583954 PDT | [0] Epoch -657 finished +---------------------------------- --------------- +epoch -657 +replay_buffer/size 999033 +trainer/num train calls 344000 +trainer/Policy Loss -20.2584 +trainer/Log Pis Mean 25.166 +trainer/Log Pis Std 12.9629 +trainer/Log Pis Max 63.7816 +trainer/Log Pis Min -6.89916 +trainer/policy/mean Mean -0.0442921 +trainer/policy/mean Std 0.906377 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.76902 +trainer/policy/normal/std Std 0.634894 +trainer/policy/normal/std Max 6.39548 +trainer/policy/normal/std Min 0.362344 +trainer/policy/normal/log_std Mean 0.979066 +trainer/policy/normal/log_std Std 0.319795 +trainer/policy/normal/log_std Max 1.85559 +trainer/policy/normal/log_std Min -1.01516 +eval/num steps total 341227 +eval/num paths total 344 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115562 +eval/Actions Std 0.907257 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82765 +time/logging (s) 0.00437366 +time/sampling batch (s) 0.292455 +time/saving (s) 0.00421146 +time/training (s) 6.51214 +time/epoch (s) 9.64084 +time/total (s) 3545.4 +Epoch -657 +---------------------------------- --------------- +2022-05-10 14:09:55.883236 PDT | [0] Epoch -656 finished +---------------------------------- --------------- +epoch -656 +replay_buffer/size 999033 +trainer/num train calls 345000 +trainer/Policy Loss -19.9617 +trainer/Log Pis Mean 25.0179 +trainer/Log Pis Std 13.1392 +trainer/Log Pis Max 74.8188 +trainer/Log Pis Min -6.90739 +trainer/policy/mean Mean -0.0253905 +trainer/policy/mean Std 0.909344 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.86563 +trainer/policy/normal/std Std 0.626495 +trainer/policy/normal/std Max 4.98981 +trainer/policy/normal/std Min 0.352035 +trainer/policy/normal/log_std Mean 1.01756 +trainer/policy/normal/log_std Std 0.300637 +trainer/policy/normal/log_std Max 1.6074 +trainer/policy/normal/log_std Min -1.04402 +eval/num steps total 342227 +eval/num paths total 345 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117752 +eval/Actions Std 0.881775 +eval/Actions Max 0.999999 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80884 +time/logging (s) 0.00391467 +time/sampling batch (s) 0.534384 +time/saving (s) 0.00371207 +time/training (s) 7.92662 +time/epoch (s) 11.2775 +time/total (s) 3556.69 +Epoch -656 +---------------------------------- --------------- +2022-05-10 14:10:05.999412 PDT | [0] Epoch -655 finished +---------------------------------- --------------- +epoch -655 +replay_buffer/size 999033 +trainer/num train calls 346000 +trainer/Policy Loss -19.5086 +trainer/Log Pis Mean 24.2672 +trainer/Log Pis Std 13.0283 +trainer/Log Pis Max 69.1769 +trainer/Log Pis Min -8.83672 +trainer/policy/mean Mean -0.0318619 +trainer/policy/mean Std 0.903426 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78184 +trainer/policy/normal/std Std 0.634729 +trainer/policy/normal/std Max 6.17646 +trainer/policy/normal/std Min 0.312975 +trainer/policy/normal/log_std Mean 0.986795 +trainer/policy/normal/log_std Std 0.300973 +trainer/policy/normal/log_std Max 1.82075 +trainer/policy/normal/log_std Min -1.16163 +eval/num steps total 343227 +eval/num paths total 346 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.141238 +eval/Actions Std 0.880046 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67394 +time/logging (s) 0.00374729 +time/sampling batch (s) 0.284372 +time/saving (s) 0.00354968 +time/training (s) 7.12952 +time/epoch (s) 10.0951 +time/total (s) 3566.78 +Epoch -655 +---------------------------------- --------------- +2022-05-10 14:10:16.245715 PDT | [0] Epoch -654 finished +---------------------------------- --------------- +epoch -654 +replay_buffer/size 999033 +trainer/num train calls 347000 +trainer/Policy Loss -18.5207 +trainer/Log Pis Mean 23.1777 +trainer/Log Pis Std 12.4482 +trainer/Log Pis Max 69.8474 +trainer/Log Pis Min -8.29923 +trainer/policy/mean Mean -0.035743 +trainer/policy/mean Std 0.903045 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76198 +trainer/policy/normal/std Std 0.609036 +trainer/policy/normal/std Max 5.48206 +trainer/policy/normal/std Min 0.348669 +trainer/policy/normal/log_std Mean 0.981866 +trainer/policy/normal/log_std Std 0.291219 +trainer/policy/normal/log_std Max 1.70148 +trainer/policy/normal/log_std Min -1.05363 +eval/num steps total 344227 +eval/num paths total 347 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108747 +eval/Actions Std 0.897813 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7066 +time/logging (s) 0.00374815 +time/sampling batch (s) 0.280887 +time/saving (s) 0.00345469 +time/training (s) 7.23103 +time/epoch (s) 10.2257 +time/total (s) 3577.01 +Epoch -654 +---------------------------------- --------------- +2022-05-10 14:10:26.392316 PDT | [0] Epoch -653 finished +---------------------------------- --------------- +epoch -653 +replay_buffer/size 999033 +trainer/num train calls 348000 +trainer/Policy Loss -19.6724 +trainer/Log Pis Mean 25.1967 +trainer/Log Pis Std 13.3316 +trainer/Log Pis Max 74.8401 +trainer/Log Pis Min -8.84235 +trainer/policy/mean Mean -0.0606907 +trainer/policy/mean Std 0.902822 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75143 +trainer/policy/normal/std Std 0.628714 +trainer/policy/normal/std Max 5.2693 +trainer/policy/normal/std Min 0.366986 +trainer/policy/normal/log_std Mean 0.973767 +trainer/policy/normal/log_std Std 0.31327 +trainer/policy/normal/log_std Max 1.6619 +trainer/policy/normal/log_std Min -1.00243 +eval/num steps total 345227 +eval/num paths total 348 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0629993 +eval/Actions Std 0.914263 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80113 +time/logging (s) 0.00381228 +time/sampling batch (s) 0.530671 +time/saving (s) 0.0035459 +time/training (s) 6.78699 +time/epoch (s) 10.1262 +time/total (s) 3587.14 +Epoch -653 +---------------------------------- --------------- +2022-05-10 14:10:35.599135 PDT | [0] Epoch -652 finished +---------------------------------- --------------- +epoch -652 +replay_buffer/size 999033 +trainer/num train calls 349000 +trainer/Policy Loss -19.3361 +trainer/Log Pis Mean 24.1943 +trainer/Log Pis Std 13.5429 +trainer/Log Pis Max 65.5223 +trainer/Log Pis Min -11.8626 +trainer/policy/mean Mean -0.0482681 +trainer/policy/mean Std 0.90781 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.82565 +trainer/policy/normal/std Std 0.627564 +trainer/policy/normal/std Max 5.17798 +trainer/policy/normal/std Min 0.353292 +trainer/policy/normal/log_std Mean 1.00307 +trainer/policy/normal/log_std Std 0.301123 +trainer/policy/normal/log_std Max 1.64441 +trainer/policy/normal/log_std Min -1.04046 +eval/num steps total 346227 +eval/num paths total 349 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.018581 +eval/Actions Std 0.904148 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51185 +time/logging (s) 0.00390543 +time/sampling batch (s) 0.279689 +time/saving (s) 0.0035522 +time/training (s) 6.38735 +time/epoch (s) 9.18635 +time/total (s) 3596.33 +Epoch -652 +---------------------------------- --------------- +2022-05-10 14:10:44.787849 PDT | [0] Epoch -651 finished +---------------------------------- --------------- +epoch -651 +replay_buffer/size 999033 +trainer/num train calls 350000 +trainer/Policy Loss -19.5642 +trainer/Log Pis Mean 24.1258 +trainer/Log Pis Std 13.0744 +trainer/Log Pis Max 66.6877 +trainer/Log Pis Min -9.03367 +trainer/policy/mean Mean -0.0287639 +trainer/policy/mean Std 0.906681 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.7949 +trainer/policy/normal/std Std 0.593332 +trainer/policy/normal/std Max 5.66697 +trainer/policy/normal/std Min 0.366006 +trainer/policy/normal/log_std Mean 0.995587 +trainer/policy/normal/log_std Std 0.28538 +trainer/policy/normal/log_std Max 1.73466 +trainer/policy/normal/log_std Min -1.0051 +eval/num steps total 347227 +eval/num paths total 350 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.248023 +eval/Actions Std 0.927948 +eval/Actions Max 0.999982 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82107 +time/logging (s) 0.00407065 +time/sampling batch (s) 0.533532 +time/saving (s) 0.00405092 +time/training (s) 5.80493 +time/epoch (s) 9.16765 +time/total (s) 3605.5 +Epoch -651 +---------------------------------- --------------- +2022-05-10 14:10:54.816772 PDT | [0] Epoch -650 finished +---------------------------------- --------------- +epoch -650 +replay_buffer/size 999033 +trainer/num train calls 351000 +trainer/Policy Loss -19.8571 +trainer/Log Pis Mean 23.9677 +trainer/Log Pis Std 13.6856 +trainer/Log Pis Max 71.7179 +trainer/Log Pis Min -6.28634 +trainer/policy/mean Mean -0.0321741 +trainer/policy/mean Std 0.906251 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.83591 +trainer/policy/normal/std Std 0.636917 +trainer/policy/normal/std Max 5.81771 +trainer/policy/normal/std Min 0.356934 +trainer/policy/normal/log_std Mean 1.0062 +trainer/policy/normal/log_std Std 0.302474 +trainer/policy/normal/log_std Max 1.76091 +trainer/policy/normal/log_std Min -1.0302 +eval/num steps total 347757 +eval/num paths total 351 +eval/path length Mean 530 +eval/path length Std 0 +eval/path length Max 530 +eval/path length Min 530 +eval/Rewards Mean 0.00188679 +eval/Rewards Std 0.0433962 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0270007 +eval/Actions Std 0.898154 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.64614 +time/logging (s) 0.0024271 +time/sampling batch (s) 0.281487 +time/saving (s) 0.00353371 +time/training (s) 7.0729 +time/epoch (s) 10.0065 +time/total (s) 3615.51 +Epoch -650 +---------------------------------- --------------- +2022-05-10 14:11:04.869194 PDT | [0] Epoch -649 finished +---------------------------------- --------------- +epoch -649 +replay_buffer/size 999033 +trainer/num train calls 352000 +trainer/Policy Loss -19.855 +trainer/Log Pis Mean 24.6733 +trainer/Log Pis Std 13.1059 +trainer/Log Pis Max 70.4452 +trainer/Log Pis Min -5.03288 +trainer/policy/mean Mean -0.0476396 +trainer/policy/mean Std 0.905539 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81602 +trainer/policy/normal/std Std 0.622334 +trainer/policy/normal/std Max 5.36763 +trainer/policy/normal/std Min 0.277645 +trainer/policy/normal/log_std Mean 0.999973 +trainer/policy/normal/log_std Std 0.300473 +trainer/policy/normal/log_std Max 1.68039 +trainer/policy/normal/log_std Min -1.28141 +eval/num steps total 348757 +eval/num paths total 352 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0584783 +eval/Actions Std 0.91419 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62753 +time/logging (s) 0.00370211 +time/sampling batch (s) 0.281045 +time/saving (s) 0.00344792 +time/training (s) 7.11746 +time/epoch (s) 10.0332 +time/total (s) 3625.55 +Epoch -649 +---------------------------------- --------------- +2022-05-10 14:11:15.068882 PDT | [0] Epoch -648 finished +---------------------------------- --------------- +epoch -648 +replay_buffer/size 999033 +trainer/num train calls 353000 +trainer/Policy Loss -18.9616 +trainer/Log Pis Mean 24.5572 +trainer/Log Pis Std 13.4226 +trainer/Log Pis Max 67.7684 +trainer/Log Pis Min -4.33068 +trainer/policy/mean Mean -0.0394683 +trainer/policy/mean Std 0.902956 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.85822 +trainer/policy/normal/std Std 0.615189 +trainer/policy/normal/std Max 6.17348 +trainer/policy/normal/std Min 0.405023 +trainer/policy/normal/log_std Mean 1.0187 +trainer/policy/normal/log_std Std 0.27775 +trainer/policy/normal/log_std Max 1.82026 +trainer/policy/normal/log_std Min -0.903812 +eval/num steps total 349757 +eval/num paths total 353 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.239971 +eval/Actions Std 0.948688 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71254 +time/logging (s) 0.00370859 +time/sampling batch (s) 0.282058 +time/saving (s) 0.00346975 +time/training (s) 7.17725 +time/epoch (s) 10.179 +time/total (s) 3635.73 +Epoch -648 +---------------------------------- --------------- +2022-05-10 14:11:24.891456 PDT | [0] Epoch -647 finished +---------------------------------- --------------- +epoch -647 +replay_buffer/size 999033 +trainer/num train calls 354000 +trainer/Policy Loss -19.7772 +trainer/Log Pis Mean 24.7145 +trainer/Log Pis Std 12.5863 +trainer/Log Pis Max 79.8341 +trainer/Log Pis Min -8.95045 +trainer/policy/mean Mean -0.0279351 +trainer/policy/mean Std 0.911425 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79374 +trainer/policy/normal/std Std 0.625861 +trainer/policy/normal/std Max 5.14 +trainer/policy/normal/std Min 0.377891 +trainer/policy/normal/log_std Mean 0.992394 +trainer/policy/normal/log_std Std 0.294038 +trainer/policy/normal/log_std Max 1.63705 +trainer/policy/normal/log_std Min -0.973149 +eval/num steps total 350757 +eval/num paths total 354 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0254659 +eval/Actions Std 0.882261 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80475 +time/logging (s) 0.00373568 +time/sampling batch (s) 0.280755 +time/saving (s) 0.00350251 +time/training (s) 6.70919 +time/epoch (s) 9.80194 +time/total (s) 3645.53 +Epoch -647 +---------------------------------- --------------- +2022-05-10 14:11:33.976522 PDT | [0] Epoch -646 finished +---------------------------------- --------------- +epoch -646 +replay_buffer/size 999033 +trainer/num train calls 355000 +trainer/Policy Loss -19.5414 +trainer/Log Pis Mean 25.0641 +trainer/Log Pis Std 13.3221 +trainer/Log Pis Max 69.0326 +trainer/Log Pis Min -6.19034 +trainer/policy/mean Mean -0.0297453 +trainer/policy/mean Std 0.90926 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8803 +trainer/policy/normal/std Std 0.644124 +trainer/policy/normal/std Max 6.02315 +trainer/policy/normal/std Min 0.339955 +trainer/policy/normal/log_std Mean 1.0221 +trainer/policy/normal/log_std Std 0.301907 +trainer/policy/normal/log_std Max 1.79561 +trainer/policy/normal/log_std Min -1.07894 +eval/num steps total 351757 +eval/num paths total 355 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0794992 +eval/Actions Std 0.904436 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5924 +time/logging (s) 0.00366445 +time/sampling batch (s) 0.280682 +time/saving (s) 0.00343442 +time/training (s) 6.18422 +time/epoch (s) 9.0644 +time/total (s) 3654.6 +Epoch -646 +---------------------------------- --------------- +2022-05-10 14:11:44.522534 PDT | [0] Epoch -645 finished +---------------------------------- --------------- +epoch -645 +replay_buffer/size 999033 +trainer/num train calls 356000 +trainer/Policy Loss -18.4563 +trainer/Log Pis Mean 23.2711 +trainer/Log Pis Std 12.9484 +trainer/Log Pis Max 69.2679 +trainer/Log Pis Min -9.58559 +trainer/policy/mean Mean -0.0443567 +trainer/policy/mean Std 0.902111 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.73458 +trainer/policy/normal/std Std 0.604008 +trainer/policy/normal/std Max 4.96367 +trainer/policy/normal/std Min 0.300213 +trainer/policy/normal/log_std Mean 0.971326 +trainer/policy/normal/log_std Std 0.295023 +trainer/policy/normal/log_std Max 1.60214 +trainer/policy/normal/log_std Min -1.20326 +eval/num steps total 352757 +eval/num paths total 356 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.192917 +eval/Actions Std 0.901614 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73519 +time/logging (s) 0.00447237 +time/sampling batch (s) 0.290348 +time/saving (s) 0.00445315 +time/training (s) 7.49139 +time/epoch (s) 10.5259 +time/total (s) 3665.13 +Epoch -645 +---------------------------------- --------------- +2022-05-10 14:11:54.132973 PDT | [0] Epoch -644 finished +---------------------------------- --------------- +epoch -644 +replay_buffer/size 999033 +trainer/num train calls 357000 +trainer/Policy Loss -19.18 +trainer/Log Pis Mean 24.3413 +trainer/Log Pis Std 14.0167 +trainer/Log Pis Max 66.8086 +trainer/Log Pis Min -7.51532 +trainer/policy/mean Mean -0.0491405 +trainer/policy/mean Std 0.907332 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82386 +trainer/policy/normal/std Std 0.608404 +trainer/policy/normal/std Max 6.20378 +trainer/policy/normal/std Min 0.378932 +trainer/policy/normal/log_std Mean 1.00694 +trainer/policy/normal/log_std Std 0.274831 +trainer/policy/normal/log_std Max 1.82516 +trainer/policy/normal/log_std Min -0.970398 +eval/num steps total 353757 +eval/num paths total 357 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.288587 +eval/Actions Std 0.899399 +eval/Actions Max 0.999995 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82947 +time/logging (s) 0.00366287 +time/sampling batch (s) 0.280325 +time/saving (s) 0.00347377 +time/training (s) 6.47171 +time/epoch (s) 9.58863 +time/total (s) 3674.72 +Epoch -644 +---------------------------------- --------------- +2022-05-10 14:12:04.110917 PDT | [0] Epoch -643 finished +---------------------------------- --------------- +epoch -643 +replay_buffer/size 999033 +trainer/num train calls 358000 +trainer/Policy Loss -20.6297 +trainer/Log Pis Mean 25.5245 +trainer/Log Pis Std 14.2331 +trainer/Log Pis Max 71.4867 +trainer/Log Pis Min -6.53188 +trainer/policy/mean Mean -0.0606431 +trainer/policy/mean Std 0.902921 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.85895 +trainer/policy/normal/std Std 0.634544 +trainer/policy/normal/std Max 5.66611 +trainer/policy/normal/std Min 0.357292 +trainer/policy/normal/log_std Mean 1.01493 +trainer/policy/normal/log_std Std 0.300956 +trainer/policy/normal/log_std Max 1.7345 +trainer/policy/normal/log_std Min -1.0292 +eval/num steps total 354757 +eval/num paths total 358 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00334157 +eval/Actions Std 0.930191 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55467 +time/logging (s) 0.00371471 +time/sampling batch (s) 0.2805 +time/saving (s) 0.00343805 +time/training (s) 7.11516 +time/epoch (s) 9.95748 +time/total (s) 3684.68 +Epoch -643 +---------------------------------- --------------- +2022-05-10 14:12:13.778625 PDT | [0] Epoch -642 finished +---------------------------------- --------------- +epoch -642 +replay_buffer/size 999033 +trainer/num train calls 359000 +trainer/Policy Loss -19.9078 +trainer/Log Pis Mean 24.0805 +trainer/Log Pis Std 13.4542 +trainer/Log Pis Max 61.8473 +trainer/Log Pis Min -4.53824 +trainer/policy/mean Mean -0.0447487 +trainer/policy/mean Std 0.908511 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81169 +trainer/policy/normal/std Std 0.608822 +trainer/policy/normal/std Max 5.32876 +trainer/policy/normal/std Min 0.315922 +trainer/policy/normal/log_std Mean 1.0008 +trainer/policy/normal/log_std Std 0.287445 +trainer/policy/normal/log_std Max 1.67312 +trainer/policy/normal/log_std Min -1.15226 +eval/num steps total 355757 +eval/num paths total 359 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.31124 +eval/Actions Std 0.883153 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74853 +time/logging (s) 0.00367399 +time/sampling batch (s) 0.279189 +time/saving (s) 0.00352839 +time/training (s) 6.61232 +time/epoch (s) 9.64724 +time/total (s) 3694.33 +Epoch -642 +---------------------------------- --------------- +2022-05-10 14:12:23.751759 PDT | [0] Epoch -641 finished +---------------------------------- --------------- +epoch -641 +replay_buffer/size 999033 +trainer/num train calls 360000 +trainer/Policy Loss -19.2919 +trainer/Log Pis Mean 24.6861 +trainer/Log Pis Std 13.521 +trainer/Log Pis Max 72.7708 +trainer/Log Pis Min -9.40006 +trainer/policy/mean Mean -0.0697128 +trainer/policy/mean Std 0.901672 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80635 +trainer/policy/normal/std Std 0.607368 +trainer/policy/normal/std Max 5.82915 +trainer/policy/normal/std Min 0.37087 +trainer/policy/normal/log_std Mean 0.999361 +trainer/policy/normal/log_std Std 0.283803 +trainer/policy/normal/log_std Max 1.76287 +trainer/policy/normal/log_std Min -0.991904 +eval/num steps total 356757 +eval/num paths total 360 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.139069 +eval/Actions Std 0.881168 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51068 +time/logging (s) 0.00366006 +time/sampling batch (s) 0.280165 +time/saving (s) 0.00344025 +time/training (s) 7.15468 +time/epoch (s) 9.95262 +time/total (s) 3704.29 +Epoch -641 +---------------------------------- --------------- +2022-05-10 14:12:33.436166 PDT | [0] Epoch -640 finished +---------------------------------- --------------- +epoch -640 +replay_buffer/size 999033 +trainer/num train calls 361000 +trainer/Policy Loss -19.9791 +trainer/Log Pis Mean 24.4405 +trainer/Log Pis Std 13.406 +trainer/Log Pis Max 82.7706 +trainer/Log Pis Min -6.50397 +trainer/policy/mean Mean -0.0270697 +trainer/policy/mean Std 0.904334 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85501 +trainer/policy/normal/std Std 0.624642 +trainer/policy/normal/std Max 5.60357 +trainer/policy/normal/std Min 0.392865 +trainer/policy/normal/log_std Mean 1.01589 +trainer/policy/normal/log_std Std 0.286676 +trainer/policy/normal/log_std Max 1.7234 +trainer/policy/normal/log_std Min -0.934291 +eval/num steps total 357757 +eval/num paths total 361 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.383976 +eval/Actions Std 0.841775 +eval/Actions Max 0.999983 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.2657 +time/logging (s) 0.00418659 +time/sampling batch (s) 0.283509 +time/saving (s) 0.0041186 +time/training (s) 7.10653 +time/epoch (s) 9.66404 +time/total (s) 3713.95 +Epoch -640 +---------------------------------- --------------- +2022-05-10 14:12:43.212375 PDT | [0] Epoch -639 finished +---------------------------------- --------------- +epoch -639 +replay_buffer/size 999033 +trainer/num train calls 362000 +trainer/Policy Loss -19.7137 +trainer/Log Pis Mean 25.138 +trainer/Log Pis Std 13.2462 +trainer/Log Pis Max 65.5547 +trainer/Log Pis Min -10.5184 +trainer/policy/mean Mean -0.0495868 +trainer/policy/mean Std 0.904115 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82709 +trainer/policy/normal/std Std 0.601084 +trainer/policy/normal/std Max 5.49711 +trainer/policy/normal/std Min 0.299505 +trainer/policy/normal/log_std Mean 1.00755 +trainer/policy/normal/log_std Std 0.281586 +trainer/policy/normal/log_std Max 1.70422 +trainer/policy/normal/log_std Min -1.20562 +eval/num steps total 358757 +eval/num paths total 362 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.208778 +eval/Actions Std 0.952892 +eval/Actions Max 0.999991 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61466 +time/logging (s) 0.003828 +time/sampling batch (s) 0.290476 +time/saving (s) 0.00388938 +time/training (s) 6.84148 +time/epoch (s) 9.75434 +time/total (s) 3723.71 +Epoch -639 +---------------------------------- --------------- +2022-05-10 14:12:52.463388 PDT | [0] Epoch -638 finished +---------------------------------- --------------- +epoch -638 +replay_buffer/size 999033 +trainer/num train calls 363000 +trainer/Policy Loss -19.7757 +trainer/Log Pis Mean 24.6246 +trainer/Log Pis Std 13.349 +trainer/Log Pis Max 69.8892 +trainer/Log Pis Min -7.50014 +trainer/policy/mean Mean -0.0420415 +trainer/policy/mean Std 0.90371 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85913 +trainer/policy/normal/std Std 0.618896 +trainer/policy/normal/std Max 5.55608 +trainer/policy/normal/std Min 0.338514 +trainer/policy/normal/log_std Mean 1.01761 +trainer/policy/normal/log_std Std 0.287369 +trainer/policy/normal/log_std Max 1.71489 +trainer/policy/normal/log_std Min -1.08319 +eval/num steps total 359385 +eval/num paths total 363 +eval/path length Mean 628 +eval/path length Std 0 +eval/path length Max 628 +eval/path length Min 628 +eval/Rewards Mean 0.00159236 +eval/Rewards Std 0.0398726 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.038678 +eval/Actions Std 0.898081 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.49265 +time/logging (s) 0.00266543 +time/sampling batch (s) 0.289605 +time/saving (s) 0.00344776 +time/training (s) 6.43994 +time/epoch (s) 9.2283 +time/total (s) 3732.94 +Epoch -638 +---------------------------------- --------------- +2022-05-10 14:13:02.332303 PDT | [0] Epoch -637 finished +---------------------------------- --------------- +epoch -637 +replay_buffer/size 999033 +trainer/num train calls 364000 +trainer/Policy Loss -19.5522 +trainer/Log Pis Mean 25.5068 +trainer/Log Pis Std 13.1492 +trainer/Log Pis Max 62.6318 +trainer/Log Pis Min -8.89377 +trainer/policy/mean Mean -0.0522927 +trainer/policy/mean Std 0.904367 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.88977 +trainer/policy/normal/std Std 0.633502 +trainer/policy/normal/std Max 5.52288 +trainer/policy/normal/std Min 0.325218 +trainer/policy/normal/log_std Mean 1.0267 +trainer/policy/normal/log_std Std 0.295891 +trainer/policy/normal/log_std Max 1.7089 +trainer/policy/normal/log_std Min -1.12326 +eval/num steps total 360385 +eval/num paths total 364 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.24066 +eval/Actions Std 0.888396 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68121 +time/logging (s) 0.00384697 +time/sampling batch (s) 0.279876 +time/saving (s) 0.00344784 +time/training (s) 6.88136 +time/epoch (s) 9.84975 +time/total (s) 3742.79 +Epoch -637 +---------------------------------- --------------- +2022-05-10 14:13:11.827213 PDT | [0] Epoch -636 finished +---------------------------------- --------------- +epoch -636 +replay_buffer/size 999033 +trainer/num train calls 365000 +trainer/Policy Loss -19.3384 +trainer/Log Pis Mean 24.349 +trainer/Log Pis Std 13.2873 +trainer/Log Pis Max 71.317 +trainer/Log Pis Min -9.33133 +trainer/policy/mean Mean -0.0240619 +trainer/policy/mean Std 0.906838 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78619 +trainer/policy/normal/std Std 0.639759 +trainer/policy/normal/std Max 6.86222 +trainer/policy/normal/std Min 0.301879 +trainer/policy/normal/log_std Mean 0.98646 +trainer/policy/normal/log_std Std 0.31245 +trainer/policy/normal/log_std Max 1.92603 +trainer/policy/normal/log_std Min -1.19773 +eval/num steps total 361385 +eval/num paths total 365 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0909987 +eval/Actions Std 0.919197 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47053 +time/logging (s) 0.0038338 +time/sampling batch (s) 0.28079 +time/saving (s) 0.00352343 +time/training (s) 6.71558 +time/epoch (s) 9.47426 +time/total (s) 3752.27 +Epoch -636 +---------------------------------- --------------- +2022-05-10 14:13:21.261539 PDT | [0] Epoch -635 finished +---------------------------------- --------------- +epoch -635 +replay_buffer/size 999033 +trainer/num train calls 366000 +trainer/Policy Loss -18.1861 +trainer/Log Pis Mean 24.42 +trainer/Log Pis Std 13.747 +trainer/Log Pis Max 82.318 +trainer/Log Pis Min -10.5656 +trainer/policy/mean Mean -0.011651 +trainer/policy/mean Std 0.907828 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81496 +trainer/policy/normal/std Std 0.639009 +trainer/policy/normal/std Max 6.05616 +trainer/policy/normal/std Min 0.34321 +trainer/policy/normal/log_std Mean 0.99896 +trainer/policy/normal/log_std Std 0.299685 +trainer/policy/normal/log_std Max 1.80108 +trainer/policy/normal/log_std Min -1.06941 +eval/num steps total 362385 +eval/num paths total 366 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0983065 +eval/Actions Std 0.917963 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65727 +time/logging (s) 0.00397085 +time/sampling batch (s) 0.287662 +time/saving (s) 0.00367988 +time/training (s) 6.46079 +time/epoch (s) 9.41337 +time/total (s) 3761.69 +Epoch -635 +---------------------------------- --------------- +2022-05-10 14:13:31.348357 PDT | [0] Epoch -634 finished +---------------------------------- --------------- +epoch -634 +replay_buffer/size 999033 +trainer/num train calls 367000 +trainer/Policy Loss -20.4011 +trainer/Log Pis Mean 24.3203 +trainer/Log Pis Std 13.1338 +trainer/Log Pis Max 79.8153 +trainer/Log Pis Min -6.39694 +trainer/policy/mean Mean -0.0375208 +trainer/policy/mean Std 0.903253 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7882 +trainer/policy/normal/std Std 0.620228 +trainer/policy/normal/std Max 6.27756 +trainer/policy/normal/std Min 0.370069 +trainer/policy/normal/log_std Mean 0.989588 +trainer/policy/normal/log_std Std 0.302117 +trainer/policy/normal/log_std Max 1.83698 +trainer/policy/normal/log_std Min -0.994065 +eval/num steps total 363385 +eval/num paths total 367 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0637201 +eval/Actions Std 0.909033 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74734 +time/logging (s) 0.00397947 +time/sampling batch (s) 0.280176 +time/saving (s) 0.00425552 +time/training (s) 7.03013 +time/epoch (s) 10.0659 +time/total (s) 3771.76 +Epoch -634 +---------------------------------- --------------- +2022-05-10 14:13:41.228170 PDT | [0] Epoch -633 finished +---------------------------------- --------------- +epoch -633 +replay_buffer/size 999033 +trainer/num train calls 368000 +trainer/Policy Loss -19.8784 +trainer/Log Pis Mean 24.3335 +trainer/Log Pis Std 12.4062 +trainer/Log Pis Max 67.2876 +trainer/Log Pis Min -4.15803 +trainer/policy/mean Mean -0.0199356 +trainer/policy/mean Std 0.904991 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83635 +trainer/policy/normal/std Std 0.630045 +trainer/policy/normal/std Max 6.48066 +trainer/policy/normal/std Min 0.353876 +trainer/policy/normal/log_std Mean 1.00773 +trainer/policy/normal/log_std Std 0.295352 +trainer/policy/normal/log_std Max 1.86882 +trainer/policy/normal/log_std Min -1.03881 +eval/num steps total 364385 +eval/num paths total 368 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.124275 +eval/Actions Std 0.910446 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48588 +time/logging (s) 0.00370045 +time/sampling batch (s) 0.282958 +time/saving (s) 0.00348991 +time/training (s) 7.08227 +time/epoch (s) 9.8583 +time/total (s) 3781.62 +Epoch -633 +---------------------------------- --------------- +2022-05-10 14:13:51.614313 PDT | [0] Epoch -632 finished +---------------------------------- --------------- +epoch -632 +replay_buffer/size 999033 +trainer/num train calls 369000 +trainer/Policy Loss -18.9919 +trainer/Log Pis Mean 24.677 +trainer/Log Pis Std 12.7331 +trainer/Log Pis Max 69.0141 +trainer/Log Pis Min -6.8326 +trainer/policy/mean Mean -0.0400589 +trainer/policy/mean Std 0.906774 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77744 +trainer/policy/normal/std Std 0.632012 +trainer/policy/normal/std Max 6.21278 +trainer/policy/normal/std Min 0.363897 +trainer/policy/normal/log_std Mean 0.983974 +trainer/policy/normal/log_std Std 0.308644 +trainer/policy/normal/log_std Max 1.82661 +trainer/policy/normal/log_std Min -1.01088 +eval/num steps total 365385 +eval/num paths total 369 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136047 +eval/Actions Std 0.856329 +eval/Actions Max 0.999997 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73793 +time/logging (s) 0.00366435 +time/sampling batch (s) 0.281966 +time/saving (s) 0.00346824 +time/training (s) 7.33832 +time/epoch (s) 10.3654 +time/total (s) 3791.99 +Epoch -632 +---------------------------------- --------------- +2022-05-10 14:14:01.786382 PDT | [0] Epoch -631 finished +---------------------------------- --------------- +epoch -631 +replay_buffer/size 999033 +trainer/num train calls 370000 +trainer/Policy Loss -19.5037 +trainer/Log Pis Mean 24.5267 +trainer/Log Pis Std 13.3758 +trainer/Log Pis Max 70.3692 +trainer/Log Pis Min -7.01578 +trainer/policy/mean Mean -0.0320873 +trainer/policy/mean Std 0.908855 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8092 +trainer/policy/normal/std Std 0.61363 +trainer/policy/normal/std Max 5.18297 +trainer/policy/normal/std Min 0.338117 +trainer/policy/normal/log_std Mean 0.997884 +trainer/policy/normal/log_std Std 0.299515 +trainer/policy/normal/log_std Max 1.64538 +trainer/policy/normal/log_std Min -1.08436 +eval/num steps total 366385 +eval/num paths total 370 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0989797 +eval/Actions Std 0.915838 +eval/Actions Max 0.999989 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.36695 +time/logging (s) 0.00391074 +time/sampling batch (s) 0.53202 +time/saving (s) 0.00348884 +time/training (s) 7.24533 +time/epoch (s) 10.1517 +time/total (s) 3802.14 +Epoch -631 +---------------------------------- --------------- +2022-05-10 14:14:11.167641 PDT | [0] Epoch -630 finished +---------------------------------- --------------- +epoch -630 +replay_buffer/size 999033 +trainer/num train calls 371000 +trainer/Policy Loss -19.7843 +trainer/Log Pis Mean 23.3683 +trainer/Log Pis Std 13.0393 +trainer/Log Pis Max 64.7281 +trainer/Log Pis Min -5.53849 +trainer/policy/mean Mean -0.028803 +trainer/policy/mean Std 0.901381 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84645 +trainer/policy/normal/std Std 0.648024 +trainer/policy/normal/std Max 6.16464 +trainer/policy/normal/std Min 0.389686 +trainer/policy/normal/log_std Mean 1.00836 +trainer/policy/normal/log_std Std 0.31016 +trainer/policy/normal/log_std Max 1.81883 +trainer/policy/normal/log_std Min -0.942414 +eval/num steps total 367385 +eval/num paths total 371 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.448029 +eval/Actions Std 0.789243 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70993 +time/logging (s) 0.00376414 +time/sampling batch (s) 0.278927 +time/saving (s) 0.00350922 +time/training (s) 6.36426 +time/epoch (s) 9.36039 +time/total (s) 3811.5 +Epoch -630 +---------------------------------- --------------- +2022-05-10 14:14:20.254216 PDT | [0] Epoch -629 finished +---------------------------------- --------------- +epoch -629 +replay_buffer/size 999033 +trainer/num train calls 372000 +trainer/Policy Loss -19.4824 +trainer/Log Pis Mean 24.0191 +trainer/Log Pis Std 13.0519 +trainer/Log Pis Max 69.0696 +trainer/Log Pis Min -14.3252 +trainer/policy/mean Mean -0.0321367 +trainer/policy/mean Std 0.904689 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80285 +trainer/policy/normal/std Std 0.649197 +trainer/policy/normal/std Max 5.25248 +trainer/policy/normal/std Min 0.299328 +trainer/policy/normal/log_std Mean 0.991614 +trainer/policy/normal/log_std Std 0.315597 +trainer/policy/normal/log_std Max 1.6587 +trainer/policy/normal/log_std Min -1.20621 +eval/num steps total 368385 +eval/num paths total 372 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10402 +eval/Actions Std 0.885867 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.28993 +time/logging (s) 0.00411545 +time/sampling batch (s) 0.280468 +time/saving (s) 0.00424384 +time/training (s) 6.48741 +time/epoch (s) 9.06617 +time/total (s) 3820.57 +Epoch -629 +---------------------------------- --------------- +2022-05-10 14:14:30.657314 PDT | [0] Epoch -628 finished +---------------------------------- --------------- +epoch -628 +replay_buffer/size 999033 +trainer/num train calls 373000 +trainer/Policy Loss -19.5485 +trainer/Log Pis Mean 23.7578 +trainer/Log Pis Std 12.8257 +trainer/Log Pis Max 65.84 +trainer/Log Pis Min -9.36021 +trainer/policy/mean Mean -0.0489038 +trainer/policy/mean Std 0.902608 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.87342 +trainer/policy/normal/std Std 0.620641 +trainer/policy/normal/std Max 5.98118 +trainer/policy/normal/std Min 0.339547 +trainer/policy/normal/log_std Mean 1.02258 +trainer/policy/normal/log_std Std 0.288046 +trainer/policy/normal/log_std Max 1.78862 +trainer/policy/normal/log_std Min -1.08014 +eval/num steps total 369385 +eval/num paths total 373 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.022621 +eval/Actions Std 0.909502 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69777 +time/logging (s) 0.00392191 +time/sampling batch (s) 0.280395 +time/saving (s) 0.00365251 +time/training (s) 7.39611 +time/epoch (s) 10.3818 +time/total (s) 3830.96 +Epoch -628 +---------------------------------- --------------- +2022-05-10 14:14:41.718904 PDT | [0] Epoch -627 finished +---------------------------------- --------------- +epoch -627 +replay_buffer/size 999033 +trainer/num train calls 374000 +trainer/Policy Loss -19.6495 +trainer/Log Pis Mean 23.7399 +trainer/Log Pis Std 13.0344 +trainer/Log Pis Max 60.1081 +trainer/Log Pis Min -8.15476 +trainer/policy/mean Mean -0.0169944 +trainer/policy/mean Std 0.907576 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80196 +trainer/policy/normal/std Std 0.634969 +trainer/policy/normal/std Max 5.75594 +trainer/policy/normal/std Min 0.372808 +trainer/policy/normal/log_std Mean 0.993109 +trainer/policy/normal/log_std Std 0.307998 +trainer/policy/normal/log_std Max 1.75023 +trainer/policy/normal/log_std Min -0.986692 +eval/num steps total 370385 +eval/num paths total 374 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.128863 +eval/Actions Std 0.892561 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53766 +time/logging (s) 0.00367186 +time/sampling batch (s) 0.537674 +time/saving (s) 0.00351243 +time/training (s) 7.95741 +time/epoch (s) 11.0399 +time/total (s) 3842 +Epoch -627 +---------------------------------- --------------- +2022-05-10 14:14:51.803018 PDT | [0] Epoch -626 finished +---------------------------------- --------------- +epoch -626 +replay_buffer/size 999033 +trainer/num train calls 375000 +trainer/Policy Loss -19.6901 +trainer/Log Pis Mean 24.9112 +trainer/Log Pis Std 12.8791 +trainer/Log Pis Max 57.8351 +trainer/Log Pis Min -8.63087 +trainer/policy/mean Mean -0.0303934 +trainer/policy/mean Std 0.906408 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85544 +trainer/policy/normal/std Std 0.616025 +trainer/policy/normal/std Max 5.09251 +trainer/policy/normal/std Min 0.378386 +trainer/policy/normal/log_std Mean 1.01648 +trainer/policy/normal/log_std Std 0.285955 +trainer/policy/normal/log_std Max 1.62777 +trainer/policy/normal/log_std Min -0.971841 +eval/num steps total 371385 +eval/num paths total 375 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0159113 +eval/Actions Std 0.903892 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60759 +time/logging (s) 0.00369336 +time/sampling batch (s) 0.283429 +time/saving (s) 0.00347726 +time/training (s) 7.16505 +time/epoch (s) 10.0632 +time/total (s) 3852.07 +Epoch -626 +---------------------------------- --------------- +2022-05-10 14:15:02.352175 PDT | [0] Epoch -625 finished +---------------------------------- --------------- +epoch -625 +replay_buffer/size 999033 +trainer/num train calls 376000 +trainer/Policy Loss -19.1777 +trainer/Log Pis Mean 24.6538 +trainer/Log Pis Std 13.9255 +trainer/Log Pis Max 70.8686 +trainer/Log Pis Min -8.68441 +trainer/policy/mean Mean -0.0389097 +trainer/policy/mean Std 0.908144 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.82833 +trainer/policy/normal/std Std 0.628347 +trainer/policy/normal/std Max 5.81495 +trainer/policy/normal/std Min 0.291742 +trainer/policy/normal/log_std Mean 1.00496 +trainer/policy/normal/log_std Std 0.294298 +trainer/policy/normal/log_std Max 1.76043 +trainer/policy/normal/log_std Min -1.23189 +eval/num steps total 372385 +eval/num paths total 376 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122299 +eval/Actions Std 0.916004 +eval/Actions Max 0.999987 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77912 +time/logging (s) 0.00375352 +time/sampling batch (s) 0.533788 +time/saving (s) 0.00348616 +time/training (s) 7.20748 +time/epoch (s) 10.5276 +time/total (s) 3862.6 +Epoch -625 +---------------------------------- --------------- +2022-05-10 14:15:11.963742 PDT | [0] Epoch -624 finished +---------------------------------- --------------- +epoch -624 +replay_buffer/size 999033 +trainer/num train calls 377000 +trainer/Policy Loss -19.854 +trainer/Log Pis Mean 24.8806 +trainer/Log Pis Std 12.5016 +trainer/Log Pis Max 71.0511 +trainer/Log Pis Min -5.66607 +trainer/policy/mean Mean -0.0199457 +trainer/policy/mean Std 0.903688 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.88525 +trainer/policy/normal/std Std 0.634577 +trainer/policy/normal/std Max 5.81367 +trainer/policy/normal/std Min 0.30413 +trainer/policy/normal/log_std Mean 1.02524 +trainer/policy/normal/log_std Std 0.294853 +trainer/policy/normal/log_std Max 1.76021 +trainer/policy/normal/log_std Min -1.1903 +eval/num steps total 373385 +eval/num paths total 377 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.1131 +eval/Actions Std 0.916303 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82357 +time/logging (s) 0.00413338 +time/sampling batch (s) 0.530312 +time/saving (s) 0.00404863 +time/training (s) 6.22914 +time/epoch (s) 9.5912 +time/total (s) 3872.19 +Epoch -624 +---------------------------------- --------------- +2022-05-10 14:15:21.246157 PDT | [0] Epoch -623 finished +---------------------------------- --------------- +epoch -623 +replay_buffer/size 999033 +trainer/num train calls 378000 +trainer/Policy Loss -19.7445 +trainer/Log Pis Mean 23.5805 +trainer/Log Pis Std 12.7329 +trainer/Log Pis Max 72.4826 +trainer/Log Pis Min -6.11475 +trainer/policy/mean Mean -0.0456755 +trainer/policy/mean Std 0.903225 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7933 +trainer/policy/normal/std Std 0.62833 +trainer/policy/normal/std Max 5.7762 +trainer/policy/normal/std Min 0.290252 +trainer/policy/normal/log_std Mean 0.991475 +trainer/policy/normal/log_std Std 0.300177 +trainer/policy/normal/log_std Max 1.75375 +trainer/policy/normal/log_std Min -1.237 +eval/num steps total 374385 +eval/num paths total 378 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.261822 +eval/Actions Std 0.819398 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63422 +time/logging (s) 0.00378316 +time/sampling batch (s) 0.279253 +time/saving (s) 0.00376832 +time/training (s) 6.34008 +time/epoch (s) 9.26111 +time/total (s) 3881.46 +Epoch -623 +---------------------------------- --------------- +2022-05-10 14:15:31.347665 PDT | [0] Epoch -622 finished +---------------------------------- --------------- +epoch -622 +replay_buffer/size 999033 +trainer/num train calls 379000 +trainer/Policy Loss -19.7705 +trainer/Log Pis Mean 23.8443 +trainer/Log Pis Std 12.5466 +trainer/Log Pis Max 69.1374 +trainer/Log Pis Min -8.14158 +trainer/policy/mean Mean -0.0259537 +trainer/policy/mean Std 0.902278 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81446 +trainer/policy/normal/std Std 0.63825 +trainer/policy/normal/std Max 5.27776 +trainer/policy/normal/std Min 0.312225 +trainer/policy/normal/log_std Mean 0.997304 +trainer/policy/normal/log_std Std 0.309152 +trainer/policy/normal/log_std Max 1.6635 +trainer/policy/normal/log_std Min -1.16403 +eval/num steps total 375385 +eval/num paths total 379 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0770487 +eval/Actions Std 0.915414 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66216 +time/logging (s) 0.00371644 +time/sampling batch (s) 0.279246 +time/saving (s) 0.00341845 +time/training (s) 7.13215 +time/epoch (s) 10.0807 +time/total (s) 3891.54 +Epoch -622 +---------------------------------- --------------- +2022-05-10 14:15:41.444650 PDT | [0] Epoch -621 finished +---------------------------------- --------------- +epoch -621 +replay_buffer/size 999033 +trainer/num train calls 380000 +trainer/Policy Loss -20.0935 +trainer/Log Pis Mean 23.7502 +trainer/Log Pis Std 13.3643 +trainer/Log Pis Max 65.1942 +trainer/Log Pis Min -9.62792 +trainer/policy/mean Mean -0.0254467 +trainer/policy/mean Std 0.903953 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.77981 +trainer/policy/normal/std Std 0.634889 +trainer/policy/normal/std Max 5.23753 +trainer/policy/normal/std Min 0.324048 +trainer/policy/normal/log_std Mean 0.984297 +trainer/policy/normal/log_std Std 0.312096 +trainer/policy/normal/log_std Max 1.65585 +trainer/policy/normal/log_std Min -1.12686 +eval/num steps total 376385 +eval/num paths total 380 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122141 +eval/Actions Std 0.912894 +eval/Actions Max 0.999988 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51491 +time/logging (s) 0.00445366 +time/sampling batch (s) 0.286701 +time/saving (s) 0.00343475 +time/training (s) 7.26712 +time/epoch (s) 10.0766 +time/total (s) 3901.62 +Epoch -621 +---------------------------------- --------------- +2022-05-10 14:15:51.905051 PDT | [0] Epoch -620 finished +---------------------------------- --------------- +epoch -620 +replay_buffer/size 999033 +trainer/num train calls 381000 +trainer/Policy Loss -20.3623 +trainer/Log Pis Mean 25.5215 +trainer/Log Pis Std 13.0355 +trainer/Log Pis Max 66.5909 +trainer/Log Pis Min -6.50153 +trainer/policy/mean Mean -0.046517 +trainer/policy/mean Std 0.909489 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.84002 +trainer/policy/normal/std Std 0.608452 +trainer/policy/normal/std Max 5.67094 +trainer/policy/normal/std Min 0.344431 +trainer/policy/normal/log_std Mean 1.01144 +trainer/policy/normal/log_std Std 0.285301 +trainer/policy/normal/log_std Max 1.73535 +trainer/policy/normal/log_std Min -1.06586 +eval/num steps total 377385 +eval/num paths total 381 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.135438 +eval/Actions Std 0.884973 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63114 +time/logging (s) 0.00379267 +time/sampling batch (s) 0.289499 +time/saving (s) 0.00394969 +time/training (s) 7.50938 +time/epoch (s) 10.4378 +time/total (s) 3912.06 +Epoch -620 +---------------------------------- --------------- +2022-05-10 14:16:01.520503 PDT | [0] Epoch -619 finished +---------------------------------- --------------- +epoch -619 +replay_buffer/size 999033 +trainer/num train calls 382000 +trainer/Policy Loss -19.0148 +trainer/Log Pis Mean 23.4448 +trainer/Log Pis Std 12.6181 +trainer/Log Pis Max 66.4059 +trainer/Log Pis Min -4.82024 +trainer/policy/mean Mean -0.0349588 +trainer/policy/mean Std 0.906681 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84527 +trainer/policy/normal/std Std 0.652081 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.341579 +trainer/policy/normal/log_std Mean 1.00982 +trainer/policy/normal/log_std Std 0.297912 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.07418 +eval/num steps total 378385 +eval/num paths total 382 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120057 +eval/Actions Std 0.890879 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59267 +time/logging (s) 0.00406731 +time/sampling batch (s) 0.283459 +time/saving (s) 0.0041002 +time/training (s) 6.70986 +time/epoch (s) 9.59415 +time/total (s) 3921.66 +Epoch -619 +---------------------------------- --------------- +2022-05-10 14:16:10.977915 PDT | [0] Epoch -618 finished +---------------------------------- --------------- +epoch -618 +replay_buffer/size 999033 +trainer/num train calls 383000 +trainer/Policy Loss -20.0954 +trainer/Log Pis Mean 23.5636 +trainer/Log Pis Std 13.2157 +trainer/Log Pis Max 70.0743 +trainer/Log Pis Min -7.17047 +trainer/policy/mean Mean -0.044332 +trainer/policy/mean Std 0.906908 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78624 +trainer/policy/normal/std Std 0.635654 +trainer/policy/normal/std Max 6.35048 +trainer/policy/normal/std Min 0.301442 +trainer/policy/normal/log_std Mean 0.98801 +trainer/policy/normal/log_std Std 0.303024 +trainer/policy/normal/log_std Max 1.84853 +trainer/policy/normal/log_std Min -1.19918 +eval/num steps total 379385 +eval/num paths total 383 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00719077 +eval/Actions Std 0.876574 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60394 +time/logging (s) 0.00376765 +time/sampling batch (s) 0.280895 +time/saving (s) 0.00370385 +time/training (s) 6.54322 +time/epoch (s) 9.43553 +time/total (s) 3931.1 +Epoch -618 +---------------------------------- --------------- +2022-05-10 14:16:20.951809 PDT | [0] Epoch -617 finished +---------------------------------- --------------- +epoch -617 +replay_buffer/size 999033 +trainer/num train calls 384000 +trainer/Policy Loss -20.7273 +trainer/Log Pis Mean 25.3157 +trainer/Log Pis Std 12.9222 +trainer/Log Pis Max 65.3752 +trainer/Log Pis Min -3.94896 +trainer/policy/mean Mean -0.0324492 +trainer/policy/mean Std 0.905118 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.87101 +trainer/policy/normal/std Std 0.651135 +trainer/policy/normal/std Max 6.05059 +trainer/policy/normal/std Min 0.322625 +trainer/policy/normal/log_std Mean 1.01614 +trainer/policy/normal/log_std Std 0.316353 +trainer/policy/normal/log_std Max 1.80015 +trainer/policy/normal/log_std Min -1.13127 +eval/num steps total 380385 +eval/num paths total 384 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0996547 +eval/Actions Std 0.916915 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5011 +time/logging (s) 0.00366236 +time/sampling batch (s) 0.279003 +time/saving (s) 0.00345115 +time/training (s) 7.16588 +time/epoch (s) 9.9531 +time/total (s) 3941.05 +Epoch -617 +---------------------------------- --------------- +2022-05-10 14:16:30.271398 PDT | [0] Epoch -616 finished +---------------------------------- --------------- +epoch -616 +replay_buffer/size 999033 +trainer/num train calls 385000 +trainer/Policy Loss -18.9641 +trainer/Log Pis Mean 25.5839 +trainer/Log Pis Std 13.9083 +trainer/Log Pis Max 71.0915 +trainer/Log Pis Min -5.97311 +trainer/policy/mean Mean -0.0435912 +trainer/policy/mean Std 0.906265 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86964 +trainer/policy/normal/std Std 0.620775 +trainer/policy/normal/std Max 5.60262 +trainer/policy/normal/std Min 0.270971 +trainer/policy/normal/log_std Mean 1.02086 +trainer/policy/normal/log_std Std 0.291056 +trainer/policy/normal/log_std Max 1.72323 +trainer/policy/normal/log_std Min -1.30574 +eval/num steps total 381385 +eval/num paths total 385 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0142196 +eval/Actions Std 0.940447 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62102 +time/logging (s) 0.00371849 +time/sampling batch (s) 0.277598 +time/saving (s) 0.00343726 +time/training (s) 6.39345 +time/epoch (s) 9.29922 +time/total (s) 3950.35 +Epoch -616 +---------------------------------- --------------- +2022-05-10 14:16:39.855347 PDT | [0] Epoch -615 finished +---------------------------------- --------------- +epoch -615 +replay_buffer/size 999033 +trainer/num train calls 386000 +trainer/Policy Loss -19.8819 +trainer/Log Pis Mean 24.6631 +trainer/Log Pis Std 13.3218 +trainer/Log Pis Max 61.6216 +trainer/Log Pis Min -6.24581 +trainer/policy/mean Mean -0.0381914 +trainer/policy/mean Std 0.909253 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83 +trainer/policy/normal/std Std 0.642476 +trainer/policy/normal/std Max 5.55101 +trainer/policy/normal/std Min 0.379217 +trainer/policy/normal/log_std Mean 1.00277 +trainer/policy/normal/log_std Std 0.308628 +trainer/policy/normal/log_std Max 1.71398 +trainer/policy/normal/log_std Min -0.969646 +eval/num steps total 382385 +eval/num paths total 386 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0521875 +eval/Actions Std 0.910477 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6244 +time/logging (s) 0.00364944 +time/sampling batch (s) 0.278351 +time/saving (s) 0.00342765 +time/training (s) 6.65361 +time/epoch (s) 9.56344 +time/total (s) 3959.92 +Epoch -615 +---------------------------------- --------------- +2022-05-10 14:16:49.674487 PDT | [0] Epoch -614 finished +---------------------------------- --------------- +epoch -614 +replay_buffer/size 999033 +trainer/num train calls 387000 +trainer/Policy Loss -19.1734 +trainer/Log Pis Mean 24.3803 +trainer/Log Pis Std 13.7207 +trainer/Log Pis Max 66.374 +trainer/Log Pis Min -9.93119 +trainer/policy/mean Mean -0.0445742 +trainer/policy/mean Std 0.903594 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82575 +trainer/policy/normal/std Std 0.620157 +trainer/policy/normal/std Max 5.25911 +trainer/policy/normal/std Min 0.398308 +trainer/policy/normal/log_std Mean 1.00477 +trainer/policy/normal/log_std Std 0.291552 +trainer/policy/normal/log_std Max 1.65996 +trainer/policy/normal/log_std Min -0.92053 +eval/num steps total 383385 +eval/num paths total 387 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.151245 +eval/Actions Std 0.868266 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5266 +time/logging (s) 0.00382309 +time/sampling batch (s) 0.783305 +time/saving (s) 0.0035073 +time/training (s) 6.48122 +time/epoch (s) 9.79846 +time/total (s) 3969.72 +Epoch -614 +---------------------------------- --------------- +2022-05-10 14:16:59.661706 PDT | [0] Epoch -613 finished +---------------------------------- --------------- +epoch -613 +replay_buffer/size 999033 +trainer/num train calls 388000 +trainer/Policy Loss -19.978 +trainer/Log Pis Mean 24.8837 +trainer/Log Pis Std 13.2194 +trainer/Log Pis Max 76.5006 +trainer/Log Pis Min -7.06832 +trainer/policy/mean Mean -0.0366952 +trainer/policy/mean Std 0.90547 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81946 +trainer/policy/normal/std Std 0.625025 +trainer/policy/normal/std Max 5.3907 +trainer/policy/normal/std Min 0.311827 +trainer/policy/normal/log_std Mean 1.00174 +trainer/policy/normal/log_std Std 0.295407 +trainer/policy/normal/log_std Max 1.68468 +trainer/policy/normal/log_std Min -1.16531 +eval/num steps total 384385 +eval/num paths total 388 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.116618 +eval/Actions Std 0.824764 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59346 +time/logging (s) 0.00420933 +time/sampling batch (s) 0.295182 +time/saving (s) 0.00416575 +time/training (s) 7.06891 +time/epoch (s) 9.96592 +time/total (s) 3979.69 +Epoch -613 +---------------------------------- --------------- +2022-05-10 14:17:09.319337 PDT | [0] Epoch -612 finished +---------------------------------- --------------- +epoch -612 +replay_buffer/size 999033 +trainer/num train calls 389000 +trainer/Policy Loss -19.6548 +trainer/Log Pis Mean 24.7564 +trainer/Log Pis Std 13.1919 +trainer/Log Pis Max 68.2185 +trainer/Log Pis Min -6.88837 +trainer/policy/mean Mean -0.00809694 +trainer/policy/mean Std 0.906989 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83609 +trainer/policy/normal/std Std 0.639136 +trainer/policy/normal/std Max 5.34709 +trainer/policy/normal/std Min 0.30265 +trainer/policy/normal/log_std Mean 1.00539 +trainer/policy/normal/log_std Std 0.308217 +trainer/policy/normal/log_std Max 1.67655 +trainer/policy/normal/log_std Min -1.19518 +eval/num steps total 385385 +eval/num paths total 389 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.275764 +eval/Actions Std 0.844128 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84124 +time/logging (s) 0.00433137 +time/sampling batch (s) 0.29171 +time/saving (s) 0.00440348 +time/training (s) 6.49412 +time/epoch (s) 9.6358 +time/total (s) 3989.33 +Epoch -612 +---------------------------------- --------------- +2022-05-10 14:17:18.496413 PDT | [0] Epoch -611 finished +---------------------------------- --------------- +epoch -611 +replay_buffer/size 999033 +trainer/num train calls 390000 +trainer/Policy Loss -19.8866 +trainer/Log Pis Mean 24.9052 +trainer/Log Pis Std 12.9381 +trainer/Log Pis Max 63.5587 +trainer/Log Pis Min -3.01514 +trainer/policy/mean Mean -0.0291185 +trainer/policy/mean Std 0.907829 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85948 +trainer/policy/normal/std Std 0.631018 +trainer/policy/normal/std Max 6.15678 +trainer/policy/normal/std Min 0.350312 +trainer/policy/normal/log_std Mean 1.01599 +trainer/policy/normal/log_std Std 0.295973 +trainer/policy/normal/log_std Max 1.81755 +trainer/policy/normal/log_std Min -1.04893 +eval/num steps total 386385 +eval/num paths total 390 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0872125 +eval/Actions Std 0.875081 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68442 +time/logging (s) 0.00430405 +time/sampling batch (s) 0.290908 +time/saving (s) 0.00441656 +time/training (s) 6.17092 +time/epoch (s) 9.15496 +time/total (s) 3998.49 +Epoch -611 +---------------------------------- --------------- +2022-05-10 14:17:28.326540 PDT | [0] Epoch -610 finished +---------------------------------- --------------- +epoch -610 +replay_buffer/size 999033 +trainer/num train calls 391000 +trainer/Policy Loss -19.5466 +trainer/Log Pis Mean 24.748 +trainer/Log Pis Std 13.9079 +trainer/Log Pis Max 70.5832 +trainer/Log Pis Min -8.76574 +trainer/policy/mean Mean -0.0329767 +trainer/policy/mean Std 0.910127 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81729 +trainer/policy/normal/std Std 0.616585 +trainer/policy/normal/std Max 6.48186 +trainer/policy/normal/std Min 0.344943 +trainer/policy/normal/log_std Mean 1.00153 +trainer/policy/normal/log_std Std 0.293764 +trainer/policy/normal/log_std Max 1.86901 +trainer/policy/normal/log_std Min -1.06438 +eval/num steps total 387385 +eval/num paths total 391 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.458497 +eval/Actions Std 0.765385 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85501 +time/logging (s) 0.00420019 +time/sampling batch (s) 0.288852 +time/saving (s) 0.00434334 +time/training (s) 6.65558 +time/epoch (s) 9.80798 +time/total (s) 4008.3 +Epoch -610 +---------------------------------- --------------- +2022-05-10 14:17:37.951289 PDT | [0] Epoch -609 finished +---------------------------------- --------------- +epoch -609 +replay_buffer/size 999033 +trainer/num train calls 392000 +trainer/Policy Loss -20.366 +trainer/Log Pis Mean 24.4023 +trainer/Log Pis Std 14.175 +trainer/Log Pis Max 67.4325 +trainer/Log Pis Min -8.39964 +trainer/policy/mean Mean -0.0327966 +trainer/policy/mean Std 0.912266 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87289 +trainer/policy/normal/std Std 0.630579 +trainer/policy/normal/std Max 6.16036 +trainer/policy/normal/std Min 0.350151 +trainer/policy/normal/log_std Mean 1.02147 +trainer/policy/normal/log_std Std 0.292265 +trainer/policy/normal/log_std Max 1.81814 +trainer/policy/normal/log_std Min -1.04939 +eval/num steps total 388385 +eval/num paths total 392 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.317917 +eval/Actions Std 0.852684 +eval/Actions Max 0.999993 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53918 +time/logging (s) 0.00416672 +time/sampling batch (s) 0.282429 +time/saving (s) 0.00431396 +time/training (s) 6.77308 +time/epoch (s) 9.60317 +time/total (s) 4017.9 +Epoch -609 +---------------------------------- --------------- +2022-05-10 14:17:48.936826 PDT | [0] Epoch -608 finished +---------------------------------- --------------- +epoch -608 +replay_buffer/size 999033 +trainer/num train calls 393000 +trainer/Policy Loss -19.2881 +trainer/Log Pis Mean 23.6738 +trainer/Log Pis Std 12.5855 +trainer/Log Pis Max 71.5559 +trainer/Log Pis Min -7.03109 +trainer/policy/mean Mean -0.0266633 +trainer/policy/mean Std 0.906021 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84057 +trainer/policy/normal/std Std 0.612982 +trainer/policy/normal/std Max 5.46384 +trainer/policy/normal/std Min 0.290155 +trainer/policy/normal/log_std Mean 1.01087 +trainer/policy/normal/log_std Std 0.289735 +trainer/policy/normal/log_std Max 1.69815 +trainer/policy/normal/log_std Min -1.23734 +eval/num steps total 389385 +eval/num paths total 393 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132951 +eval/Actions Std 0.874838 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74297 +time/logging (s) 0.00378483 +time/sampling batch (s) 0.287701 +time/saving (s) 0.00358774 +time/training (s) 7.92515 +time/epoch (s) 10.9632 +time/total (s) 4028.87 +Epoch -608 +---------------------------------- --------------- +2022-05-10 14:17:58.771094 PDT | [0] Epoch -607 finished +---------------------------------- ---------------- +epoch -607 +replay_buffer/size 999033 +trainer/num train calls 394000 +trainer/Policy Loss -22.1268 +trainer/Log Pis Mean 23.9831 +trainer/Log Pis Std 12.9973 +trainer/Log Pis Max 70.9911 +trainer/Log Pis Min -4.72021 +trainer/policy/mean Mean -0.044073 +trainer/policy/mean Std 0.908577 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8402 +trainer/policy/normal/std Std 0.619898 +trainer/policy/normal/std Max 6.48757 +trainer/policy/normal/std Min 0.363579 +trainer/policy/normal/log_std Mean 1.011 +trainer/policy/normal/log_std Std 0.285683 +trainer/policy/normal/log_std Max 1.86989 +trainer/policy/normal/log_std Min -1.01176 +eval/num steps total 390385 +eval/num paths total 394 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000208637 +eval/Actions Std 0.978423 +eval/Actions Max 0.999996 +eval/Actions Min -0.999963 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50626 +time/logging (s) 0.00364033 +time/sampling batch (s) 0.280955 +time/saving (s) 0.00341119 +time/training (s) 7.01879 +time/epoch (s) 9.81306 +time/total (s) 4038.69 +Epoch -607 +---------------------------------- ---------------- +2022-05-10 14:18:08.428996 PDT | [0] Epoch -606 finished +---------------------------------- --------------- +epoch -606 +replay_buffer/size 999033 +trainer/num train calls 395000 +trainer/Policy Loss -18.639 +trainer/Log Pis Mean 25.2399 +trainer/Log Pis Std 13.112 +trainer/Log Pis Max 64.2283 +trainer/Log Pis Min -6.82186 +trainer/policy/mean Mean -0.0479915 +trainer/policy/mean Std 0.902368 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85137 +trainer/policy/normal/std Std 0.645687 +trainer/policy/normal/std Max 5.18455 +trainer/policy/normal/std Min 0.32424 +trainer/policy/normal/log_std Mean 1.01045 +trainer/policy/normal/log_std Std 0.30858 +trainer/policy/normal/log_std Max 1.64568 +trainer/policy/normal/log_std Min -1.12627 +eval/num steps total 391385 +eval/num paths total 395 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109669 +eval/Actions Std 0.907002 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58139 +time/logging (s) 0.00414558 +time/sampling batch (s) 0.279415 +time/saving (s) 0.00426899 +time/training (s) 6.76845 +time/epoch (s) 9.63767 +time/total (s) 4048.33 +Epoch -606 +---------------------------------- --------------- +2022-05-10 14:18:18.039117 PDT | [0] Epoch -605 finished +---------------------------------- --------------- +epoch -605 +replay_buffer/size 999033 +trainer/num train calls 396000 +trainer/Policy Loss -19.11 +trainer/Log Pis Mean 24.9811 +trainer/Log Pis Std 13.9345 +trainer/Log Pis Max 66.5844 +trainer/Log Pis Min -9.59116 +trainer/policy/mean Mean -0.0485357 +trainer/policy/mean Std 0.910479 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86954 +trainer/policy/normal/std Std 0.631852 +trainer/policy/normal/std Max 5.23066 +trainer/policy/normal/std Min 0.327331 +trainer/policy/normal/log_std Mean 1.01985 +trainer/policy/normal/log_std Std 0.29359 +trainer/policy/normal/log_std Max 1.65454 +trainer/policy/normal/log_std Min -1.11678 +eval/num steps total 392385 +eval/num paths total 396 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127553 +eval/Actions Std 0.899866 +eval/Actions Max 0.999983 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68513 +time/logging (s) 0.00378485 +time/sampling batch (s) 0.279785 +time/saving (s) 0.0036625 +time/training (s) 6.61581 +time/epoch (s) 9.58818 +time/total (s) 4057.92 +Epoch -605 +---------------------------------- --------------- +2022-05-10 14:18:27.558716 PDT | [0] Epoch -604 finished +---------------------------------- --------------- +epoch -604 +replay_buffer/size 999033 +trainer/num train calls 397000 +trainer/Policy Loss -20.6597 +trainer/Log Pis Mean 24.5397 +trainer/Log Pis Std 13.7751 +trainer/Log Pis Max 78.0708 +trainer/Log Pis Min -11.0392 +trainer/policy/mean Mean -0.0347949 +trainer/policy/mean Std 0.909352 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83383 +trainer/policy/normal/std Std 0.629951 +trainer/policy/normal/std Max 6.10483 +trainer/policy/normal/std Min 0.317232 +trainer/policy/normal/log_std Mean 1.00648 +trainer/policy/normal/log_std Std 0.299125 +trainer/policy/normal/log_std Max 1.80908 +trainer/policy/normal/log_std Min -1.14812 +eval/num steps total 393385 +eval/num paths total 397 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0936363 +eval/Actions Std 0.898311 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66813 +time/logging (s) 0.00379445 +time/sampling batch (s) 0.277774 +time/saving (s) 0.00345835 +time/training (s) 6.54583 +time/epoch (s) 9.49899 +time/total (s) 4067.42 +Epoch -604 +---------------------------------- --------------- +2022-05-10 14:18:37.262529 PDT | [0] Epoch -603 finished +---------------------------------- --------------- +epoch -603 +replay_buffer/size 999033 +trainer/num train calls 398000 +trainer/Policy Loss -20.2668 +trainer/Log Pis Mean 24.5765 +trainer/Log Pis Std 12.909 +trainer/Log Pis Max 62.2706 +trainer/Log Pis Min -9.99864 +trainer/policy/mean Mean -0.0520718 +trainer/policy/mean Std 0.905045 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79407 +trainer/policy/normal/std Std 0.63814 +trainer/policy/normal/std Max 5.39399 +trainer/policy/normal/std Min 0.306302 +trainer/policy/normal/log_std Mean 0.989216 +trainer/policy/normal/log_std Std 0.312977 +trainer/policy/normal/log_std Max 1.68529 +trainer/policy/normal/log_std Min -1.18318 +eval/num steps total 394385 +eval/num paths total 398 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.32667 +eval/Actions Std 0.874576 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40498 +time/logging (s) 0.00381487 +time/sampling batch (s) 0.530486 +time/saving (s) 0.00344698 +time/training (s) 6.74037 +time/epoch (s) 9.6831 +time/total (s) 4077.11 +Epoch -603 +---------------------------------- --------------- +2022-05-10 14:18:48.374907 PDT | [0] Epoch -602 finished +---------------------------------- --------------- +epoch -602 +replay_buffer/size 999033 +trainer/num train calls 399000 +trainer/Policy Loss -18.7977 +trainer/Log Pis Mean 25.5743 +trainer/Log Pis Std 13.387 +trainer/Log Pis Max 71.7503 +trainer/Log Pis Min -7.71015 +trainer/policy/mean Mean -0.0471776 +trainer/policy/mean Std 0.904151 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.89729 +trainer/policy/normal/std Std 0.638105 +trainer/policy/normal/std Max 6.13225 +trainer/policy/normal/std Min 0.317353 +trainer/policy/normal/log_std Mean 1.02992 +trainer/policy/normal/log_std Std 0.290572 +trainer/policy/normal/log_std Max 1.81356 +trainer/policy/normal/log_std Min -1.14774 +eval/num steps total 395385 +eval/num paths total 399 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00643438 +eval/Actions Std 0.904338 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59314 +time/logging (s) 0.00367359 +time/sampling batch (s) 0.53131 +time/saving (s) 0.00346516 +time/training (s) 7.95991 +time/epoch (s) 11.0915 +time/total (s) 4088.2 +Epoch -602 +---------------------------------- --------------- +2022-05-10 14:18:59.207959 PDT | [0] Epoch -601 finished +---------------------------------- -------------- +epoch -601 +replay_buffer/size 999033 +trainer/num train calls 400000 +trainer/Policy Loss -19.119 +trainer/Log Pis Mean 23.3406 +trainer/Log Pis Std 12.7677 +trainer/Log Pis Max 59.0312 +trainer/Log Pis Min -9.05391 +trainer/policy/mean Mean -0.03899 +trainer/policy/mean Std 0.906653 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75903 +trainer/policy/normal/std Std 0.614743 +trainer/policy/normal/std Max 6.80564 +trainer/policy/normal/std Min 0.298181 +trainer/policy/normal/log_std Mean 0.98014 +trainer/policy/normal/log_std Std 0.294574 +trainer/policy/normal/log_std Max 1.91775 +trainer/policy/normal/log_std Min -1.21005 +eval/num steps total 396385 +eval/num paths total 400 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.20258 +eval/Actions Std 0.932038 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65163 +time/logging (s) 0.004154 +time/sampling batch (s) 0.530665 +time/saving (s) 0.0072456 +time/training (s) 7.61894 +time/epoch (s) 10.8126 +time/total (s) 4099.02 +Epoch -601 +---------------------------------- -------------- +2022-05-10 14:19:09.543994 PDT | [0] Epoch -600 finished +---------------------------------- --------------- +epoch -600 +replay_buffer/size 999033 +trainer/num train calls 401000 +trainer/Policy Loss -19.6687 +trainer/Log Pis Mean 24.9391 +trainer/Log Pis Std 13.3943 +trainer/Log Pis Max 82.4552 +trainer/Log Pis Min -8.15146 +trainer/policy/mean Mean -0.0446631 +trainer/policy/mean Std 0.909818 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83479 +trainer/policy/normal/std Std 0.631795 +trainer/policy/normal/std Max 5.65664 +trainer/policy/normal/std Min 0.288363 +trainer/policy/normal/log_std Mean 1.00696 +trainer/policy/normal/log_std Std 0.296599 +trainer/policy/normal/log_std Max 1.73283 +trainer/policy/normal/log_std Min -1.24354 +eval/num steps total 397385 +eval/num paths total 401 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138012 +eval/Actions Std 0.881675 +eval/Actions Max 0.999988 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57776 +time/logging (s) 0.0037524 +time/sampling batch (s) 0.279884 +time/saving (s) 0.00361747 +time/training (s) 7.44918 +time/epoch (s) 10.3142 +time/total (s) 4109.33 +Epoch -600 +---------------------------------- --------------- +2022-05-10 14:19:20.658407 PDT | [0] Epoch -599 finished +---------------------------------- --------------- +epoch -599 +replay_buffer/size 999033 +trainer/num train calls 402000 +trainer/Policy Loss -19.4007 +trainer/Log Pis Mean 24.1758 +trainer/Log Pis Std 13.4438 +trainer/Log Pis Max 66.4438 +trainer/Log Pis Min -9.45734 +trainer/policy/mean Mean -0.0498401 +trainer/policy/mean Std 0.907972 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83922 +trainer/policy/normal/std Std 0.60875 +trainer/policy/normal/std Max 5.43889 +trainer/policy/normal/std Min 0.312714 +trainer/policy/normal/log_std Mean 1.01146 +trainer/policy/normal/log_std Std 0.283067 +trainer/policy/normal/log_std Max 1.69357 +trainer/policy/normal/log_std Min -1.16247 +eval/num steps total 398385 +eval/num paths total 402 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0677673 +eval/Actions Std 0.91766 +eval/Actions Max 0.999989 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61561 +time/logging (s) 0.00372059 +time/sampling batch (s) 0.28013 +time/saving (s) 0.00347735 +time/training (s) 8.19054 +time/epoch (s) 11.0935 +time/total (s) 4120.43 +Epoch -599 +---------------------------------- --------------- +2022-05-10 14:19:31.077574 PDT | [0] Epoch -598 finished +---------------------------------- --------------- +epoch -598 +replay_buffer/size 999033 +trainer/num train calls 403000 +trainer/Policy Loss -19.5901 +trainer/Log Pis Mean 24.4625 +trainer/Log Pis Std 14.0255 +trainer/Log Pis Max 77.5841 +trainer/Log Pis Min -9.56114 +trainer/policy/mean Mean -0.0527187 +trainer/policy/mean Std 0.907089 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.86332 +trainer/policy/normal/std Std 0.625276 +trainer/policy/normal/std Max 5.06794 +trainer/policy/normal/std Min 0.336356 +trainer/policy/normal/log_std Mean 1.01795 +trainer/policy/normal/log_std Std 0.293397 +trainer/policy/normal/log_std Max 1.62293 +trainer/policy/normal/log_std Min -1.08958 +eval/num steps total 399385 +eval/num paths total 403 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.289289 +eval/Actions Std 0.91203 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66782 +time/logging (s) 0.00372573 +time/sampling batch (s) 0.280342 +time/saving (s) 0.00342345 +time/training (s) 7.44325 +time/epoch (s) 10.3986 +time/total (s) 4130.83 +Epoch -598 +---------------------------------- --------------- +2022-05-10 14:19:41.124079 PDT | [0] Epoch -597 finished +---------------------------------- --------------- +epoch -597 +replay_buffer/size 999033 +trainer/num train calls 404000 +trainer/Policy Loss -18.9318 +trainer/Log Pis Mean 22.6632 +trainer/Log Pis Std 12.6633 +trainer/Log Pis Max 60.5732 +trainer/Log Pis Min -7.23885 +trainer/policy/mean Mean -0.02491 +trainer/policy/mean Std 0.90253 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81993 +trainer/policy/normal/std Std 0.621541 +trainer/policy/normal/std Max 5.43009 +trainer/policy/normal/std Min 0.354827 +trainer/policy/normal/log_std Mean 1.00182 +trainer/policy/normal/log_std Std 0.2975 +trainer/policy/normal/log_std Max 1.69195 +trainer/policy/normal/log_std Min -1.03613 +eval/num steps total 400385 +eval/num paths total 404 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0108498 +eval/Actions Std 0.909061 +eval/Actions Max 0.999998 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64665 +time/logging (s) 0.00366609 +time/sampling batch (s) 0.280146 +time/saving (s) 0.00347214 +time/training (s) 7.09161 +time/epoch (s) 10.0255 +time/total (s) 4140.86 +Epoch -597 +---------------------------------- --------------- +2022-05-10 14:19:51.287838 PDT | [0] Epoch -596 finished +---------------------------------- --------------- +epoch -596 +replay_buffer/size 999033 +trainer/num train calls 405000 +trainer/Policy Loss -20.4 +trainer/Log Pis Mean 24.6106 +trainer/Log Pis Std 13.7976 +trainer/Log Pis Max 70.9299 +trainer/Log Pis Min -13.0792 +trainer/policy/mean Mean -0.0128341 +trainer/policy/mean Std 0.902007 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.90626 +trainer/policy/normal/std Std 0.622417 +trainer/policy/normal/std Max 6.21754 +trainer/policy/normal/std Min 0.386158 +trainer/policy/normal/log_std Mean 1.03446 +trainer/policy/normal/log_std Std 0.286702 +trainer/policy/normal/log_std Max 1.82737 +trainer/policy/normal/log_std Min -0.951507 +eval/num steps total 401385 +eval/num paths total 405 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0922336 +eval/Actions Std 0.911084 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72436 +time/logging (s) 0.00411025 +time/sampling batch (s) 0.281216 +time/saving (s) 0.00407638 +time/training (s) 7.12941 +time/epoch (s) 10.1432 +time/total (s) 4151.01 +Epoch -596 +---------------------------------- --------------- +2022-05-10 14:20:01.868001 PDT | [0] Epoch -595 finished +---------------------------------- --------------- +epoch -595 +replay_buffer/size 999033 +trainer/num train calls 406000 +trainer/Policy Loss -20.0835 +trainer/Log Pis Mean 24.4718 +trainer/Log Pis Std 13.3325 +trainer/Log Pis Max 78.0846 +trainer/Log Pis Min -8.44086 +trainer/policy/mean Mean -0.0655068 +trainer/policy/mean Std 0.902147 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81907 +trainer/policy/normal/std Std 0.650088 +trainer/policy/normal/std Max 5.65331 +trainer/policy/normal/std Min 0.311575 +trainer/policy/normal/log_std Mean 0.996574 +trainer/policy/normal/log_std Std 0.321602 +trainer/policy/normal/log_std Max 1.73224 +trainer/policy/normal/log_std Min -1.16612 +eval/num steps total 402385 +eval/num paths total 406 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.430297 +eval/Actions Std 0.82888 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45224 +time/logging (s) 0.00378615 +time/sampling batch (s) 0.280578 +time/saving (s) 0.00365577 +time/training (s) 7.81839 +time/epoch (s) 10.5587 +time/total (s) 4161.57 +Epoch -595 +---------------------------------- --------------- +2022-05-10 14:20:12.711542 PDT | [0] Epoch -594 finished +---------------------------------- --------------- +epoch -594 +replay_buffer/size 999033 +trainer/num train calls 407000 +trainer/Policy Loss -19.8977 +trainer/Log Pis Mean 24.679 +trainer/Log Pis Std 13.9794 +trainer/Log Pis Max 72.9727 +trainer/Log Pis Min -11.0256 +trainer/policy/mean Mean -0.0260839 +trainer/policy/mean Std 0.908238 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83651 +trainer/policy/normal/std Std 0.629038 +trainer/policy/normal/std Max 5.64611 +trainer/policy/normal/std Min 0.361009 +trainer/policy/normal/log_std Mean 1.00713 +trainer/policy/normal/log_std Std 0.300195 +trainer/policy/normal/log_std Max 1.73097 +trainer/policy/normal/log_std Min -1.01885 +eval/num steps total 403385 +eval/num paths total 407 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.390478 +eval/Actions Std 0.894246 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73127 +time/logging (s) 0.0037719 +time/sampling batch (s) 0.281966 +time/saving (s) 0.00344502 +time/training (s) 7.80187 +time/epoch (s) 10.8223 +time/total (s) 4172.39 +Epoch -594 +---------------------------------- --------------- +2022-05-10 14:20:25.330666 PDT | [0] Epoch -593 finished +---------------------------------- --------------- +epoch -593 +replay_buffer/size 999033 +trainer/num train calls 408000 +trainer/Policy Loss -18.5854 +trainer/Log Pis Mean 24.7836 +trainer/Log Pis Std 13.7703 +trainer/Log Pis Max 76.3041 +trainer/Log Pis Min -7.97427 +trainer/policy/mean Mean -0.0516676 +trainer/policy/mean Std 0.906298 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79165 +trainer/policy/normal/std Std 0.630157 +trainer/policy/normal/std Max 5.7819 +trainer/policy/normal/std Min 0.332652 +trainer/policy/normal/log_std Mean 0.991451 +trainer/policy/normal/log_std Std 0.294781 +trainer/policy/normal/log_std Max 1.75473 +trainer/policy/normal/log_std Min -1.10066 +eval/num steps total 404385 +eval/num paths total 408 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0882009 +eval/Actions Std 0.917029 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72628 +time/logging (s) 0.00373896 +time/sampling batch (s) 0.284404 +time/saving (s) 0.00341794 +time/training (s) 9.58012 +time/epoch (s) 12.598 +time/total (s) 4184.99 +Epoch -593 +---------------------------------- --------------- +2022-05-10 14:20:35.437097 PDT | [0] Epoch -592 finished +---------------------------------- --------------- +epoch -592 +replay_buffer/size 999033 +trainer/num train calls 409000 +trainer/Policy Loss -17.7662 +trainer/Log Pis Mean 22.2228 +trainer/Log Pis Std 12.7774 +trainer/Log Pis Max 62.9803 +trainer/Log Pis Min -9.54631 +trainer/policy/mean Mean -0.0348619 +trainer/policy/mean Std 0.894973 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78844 +trainer/policy/normal/std Std 0.626641 +trainer/policy/normal/std Max 5.17902 +trainer/policy/normal/std Min 0.238567 +trainer/policy/normal/log_std Mean 0.989496 +trainer/policy/normal/log_std Std 0.301367 +trainer/policy/normal/log_std Max 1.64462 +trainer/policy/normal/log_std Min -1.43311 +eval/num steps total 405385 +eval/num paths total 409 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.328648 +eval/Actions Std 0.888865 +eval/Actions Max 0.999999 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58095 +time/logging (s) 0.00382373 +time/sampling batch (s) 0.539144 +time/saving (s) 0.00358169 +time/training (s) 6.95631 +time/epoch (s) 10.0838 +time/total (s) 4195.08 +Epoch -592 +---------------------------------- --------------- +2022-05-10 14:20:46.220513 PDT | [0] Epoch -591 finished +---------------------------------- --------------- +epoch -591 +replay_buffer/size 999033 +trainer/num train calls 410000 +trainer/Policy Loss -19.0277 +trainer/Log Pis Mean 24.7615 +trainer/Log Pis Std 13.7053 +trainer/Log Pis Max 64.2374 +trainer/Log Pis Min -6.94836 +trainer/policy/mean Mean -0.0521112 +trainer/policy/mean Std 0.904452 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85681 +trainer/policy/normal/std Std 0.641773 +trainer/policy/normal/std Max 5.64372 +trainer/policy/normal/std Min 0.346257 +trainer/policy/normal/log_std Mean 1.01404 +trainer/policy/normal/log_std Std 0.298912 +trainer/policy/normal/log_std Max 1.73054 +trainer/policy/normal/log_std Min -1.06057 +eval/num steps total 406385 +eval/num paths total 410 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.208678 +eval/Actions Std 0.903765 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65036 +time/logging (s) 0.00409432 +time/sampling batch (s) 0.29859 +time/saving (s) 0.00411271 +time/training (s) 7.80455 +time/epoch (s) 10.7617 +time/total (s) 4205.85 +Epoch -591 +---------------------------------- --------------- +2022-05-10 14:20:57.547659 PDT | [0] Epoch -590 finished +---------------------------------- --------------- +epoch -590 +replay_buffer/size 999033 +trainer/num train calls 411000 +trainer/Policy Loss -19.9471 +trainer/Log Pis Mean 24.3457 +trainer/Log Pis Std 12.9305 +trainer/Log Pis Max 65.6522 +trainer/Log Pis Min -8.13359 +trainer/policy/mean Mean -0.0385547 +trainer/policy/mean Std 0.90707 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85047 +trainer/policy/normal/std Std 0.632828 +trainer/policy/normal/std Max 6.2717 +trainer/policy/normal/std Min 0.349321 +trainer/policy/normal/log_std Mean 1.01293 +trainer/policy/normal/log_std Std 0.294266 +trainer/policy/normal/log_std Max 1.83605 +trainer/policy/normal/log_std Min -1.05177 +eval/num steps total 407385 +eval/num paths total 411 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.356598 +eval/Actions Std 0.784972 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80551 +time/logging (s) 0.00432682 +time/sampling batch (s) 0.306658 +time/saving (s) 0.00445178 +time/training (s) 8.18471 +time/epoch (s) 11.3057 +time/total (s) 4217.16 +Epoch -590 +---------------------------------- --------------- +2022-05-10 14:21:08.253072 PDT | [0] Epoch -589 finished +---------------------------------- --------------- +epoch -589 +replay_buffer/size 999033 +trainer/num train calls 412000 +trainer/Policy Loss -20.1081 +trainer/Log Pis Mean 23.8336 +trainer/Log Pis Std 13.433 +trainer/Log Pis Max 70.3018 +trainer/Log Pis Min -11.2612 +trainer/policy/mean Mean -0.0341697 +trainer/policy/mean Std 0.904382 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84691 +trainer/policy/normal/std Std 0.617325 +trainer/policy/normal/std Max 5.23821 +trainer/policy/normal/std Min 0.289806 +trainer/policy/normal/log_std Mean 1.01323 +trainer/policy/normal/log_std Std 0.287746 +trainer/policy/normal/log_std Max 1.65598 +trainer/policy/normal/log_std Min -1.23854 +eval/num steps total 408385 +eval/num paths total 412 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0606863 +eval/Actions Std 0.88128 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58621 +time/logging (s) 0.0039985 +time/sampling batch (s) 0.531494 +time/saving (s) 0.00410872 +time/training (s) 7.55733 +time/epoch (s) 10.6831 +time/total (s) 4227.84 +Epoch -589 +---------------------------------- --------------- +2022-05-10 14:21:18.873132 PDT | [0] Epoch -588 finished +---------------------------------- --------------- +epoch -588 +replay_buffer/size 999033 +trainer/num train calls 413000 +trainer/Policy Loss -18.8911 +trainer/Log Pis Mean 24.4669 +trainer/Log Pis Std 12.5385 +trainer/Log Pis Max 65.9993 +trainer/Log Pis Min -3.30871 +trainer/policy/mean Mean -0.0320983 +trainer/policy/mean Std 0.903914 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79456 +trainer/policy/normal/std Std 0.593078 +trainer/policy/normal/std Max 5.38664 +trainer/policy/normal/std Min 0.337052 +trainer/policy/normal/log_std Mean 0.996201 +trainer/policy/normal/log_std Std 0.280278 +trainer/policy/normal/log_std Max 1.68392 +trainer/policy/normal/log_std Min -1.08752 +eval/num steps total 409385 +eval/num paths total 413 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0433823 +eval/Actions Std 0.910652 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61456 +time/logging (s) 0.00407058 +time/sampling batch (s) 0.282083 +time/saving (s) 0.00413223 +time/training (s) 7.69402 +time/epoch (s) 10.5989 +time/total (s) 4238.44 +Epoch -588 +---------------------------------- --------------- +2022-05-10 14:21:30.055211 PDT | [0] Epoch -587 finished +---------------------------------- --------------- +epoch -587 +replay_buffer/size 999033 +trainer/num train calls 414000 +trainer/Policy Loss -19.1163 +trainer/Log Pis Mean 25.6374 +trainer/Log Pis Std 12.7742 +trainer/Log Pis Max 62.9501 +trainer/Log Pis Min -6.82763 +trainer/policy/mean Mean -0.0404937 +trainer/policy/mean Std 0.907727 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84735 +trainer/policy/normal/std Std 0.645762 +trainer/policy/normal/std Max 6.1913 +trainer/policy/normal/std Min 0.343527 +trainer/policy/normal/log_std Mean 1.00946 +trainer/policy/normal/log_std Std 0.306313 +trainer/policy/normal/log_std Max 1.82315 +trainer/policy/normal/log_std Min -1.06849 +eval/num steps total 410385 +eval/num paths total 414 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.137866 +eval/Actions Std 0.823669 +eval/Actions Max 0.999995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67421 +time/logging (s) 0.00418574 +time/sampling batch (s) 0.281694 +time/saving (s) 0.00413162 +time/training (s) 8.19671 +time/epoch (s) 11.1609 +time/total (s) 4249.61 +Epoch -587 +---------------------------------- --------------- +2022-05-10 14:21:40.142208 PDT | [0] Epoch -586 finished +---------------------------------- --------------- +epoch -586 +replay_buffer/size 999033 +trainer/num train calls 415000 +trainer/Policy Loss -19.3581 +trainer/Log Pis Mean 24.215 +trainer/Log Pis Std 12.5968 +trainer/Log Pis Max 61.0241 +trainer/Log Pis Min -9.88233 +trainer/policy/mean Mean -0.0417019 +trainer/policy/mean Std 0.903891 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.78301 +trainer/policy/normal/std Std 0.639456 +trainer/policy/normal/std Max 6.83163 +trainer/policy/normal/std Min 0.323571 +trainer/policy/normal/log_std Mean 0.985507 +trainer/policy/normal/log_std Std 0.310424 +trainer/policy/normal/log_std Max 1.92156 +trainer/policy/normal/log_std Min -1.12834 +eval/num steps total 411385 +eval/num paths total 415 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0945388 +eval/Actions Std 0.931093 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79912 +time/logging (s) 0.00373628 +time/sampling batch (s) 0.280503 +time/saving (s) 0.00364325 +time/training (s) 6.97808 +time/epoch (s) 10.0651 +time/total (s) 4259.68 +Epoch -586 +---------------------------------- --------------- +2022-05-10 14:21:49.890307 PDT | [0] Epoch -585 finished +---------------------------------- --------------- +epoch -585 +replay_buffer/size 999033 +trainer/num train calls 416000 +trainer/Policy Loss -19.6859 +trainer/Log Pis Mean 25.4944 +trainer/Log Pis Std 13.2495 +trainer/Log Pis Max 60.9229 +trainer/Log Pis Min -7.01501 +trainer/policy/mean Mean -0.0189067 +trainer/policy/mean Std 0.907485 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83865 +trainer/policy/normal/std Std 0.653802 +trainer/policy/normal/std Max 6.8469 +trainer/policy/normal/std Min 0.380082 +trainer/policy/normal/log_std Mean 1.00637 +trainer/policy/normal/log_std Std 0.304587 +trainer/policy/normal/log_std Max 1.9238 +trainer/policy/normal/log_std Min -0.967369 +eval/num steps total 412385 +eval/num paths total 416 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.134799 +eval/Actions Std 0.918162 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50327 +time/logging (s) 0.00365452 +time/sampling batch (s) 0.287702 +time/saving (s) 0.00348123 +time/training (s) 6.92858 +time/epoch (s) 9.72669 +time/total (s) 4269.41 +Epoch -585 +---------------------------------- --------------- +2022-05-10 14:22:00.910833 PDT | [0] Epoch -584 finished +---------------------------------- --------------- +epoch -584 +replay_buffer/size 999033 +trainer/num train calls 417000 +trainer/Policy Loss -17.3168 +trainer/Log Pis Mean 23.9752 +trainer/Log Pis Std 13.5823 +trainer/Log Pis Max 82.5383 +trainer/Log Pis Min -8.79613 +trainer/policy/mean Mean -0.0206032 +trainer/policy/mean Std 0.902114 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78487 +trainer/policy/normal/std Std 0.633844 +trainer/policy/normal/std Max 5.76854 +trainer/policy/normal/std Min 0.365831 +trainer/policy/normal/log_std Mean 0.988317 +trainer/policy/normal/log_std Std 0.298514 +trainer/policy/normal/log_std Max 1.75242 +trainer/policy/normal/log_std Min -1.00558 +eval/num steps total 413385 +eval/num paths total 417 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.260857 +eval/Actions Std 0.919228 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71359 +time/logging (s) 0.00399534 +time/sampling batch (s) 0.536702 +time/saving (s) 0.00415181 +time/training (s) 7.74105 +time/epoch (s) 10.9995 +time/total (s) 4280.41 +Epoch -584 +---------------------------------- --------------- +2022-05-10 14:22:12.862073 PDT | [0] Epoch -583 finished +---------------------------------- --------------- +epoch -583 +replay_buffer/size 999033 +trainer/num train calls 418000 +trainer/Policy Loss -20.1509 +trainer/Log Pis Mean 25.009 +trainer/Log Pis Std 13.2493 +trainer/Log Pis Max 64.0332 +trainer/Log Pis Min -5.35092 +trainer/policy/mean Mean -0.0265345 +trainer/policy/mean Std 0.911158 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85589 +trainer/policy/normal/std Std 0.605325 +trainer/policy/normal/std Max 5.39312 +trainer/policy/normal/std Min 0.386582 +trainer/policy/normal/log_std Mean 1.01711 +trainer/policy/normal/log_std Std 0.285222 +trainer/policy/normal/log_std Max 1.68512 +trainer/policy/normal/log_std Min -0.950412 +eval/num steps total 414385 +eval/num paths total 418 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.136689 +eval/Actions Std 0.909234 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54079 +time/logging (s) 0.00374658 +time/sampling batch (s) 0.536323 +time/saving (s) 0.00342594 +time/training (s) 8.84532 +time/epoch (s) 11.9296 +time/total (s) 4292.34 +Epoch -583 +---------------------------------- --------------- +2022-05-10 14:22:23.237906 PDT | [0] Epoch -582 finished +---------------------------------- --------------- +epoch -582 +replay_buffer/size 999033 +trainer/num train calls 419000 +trainer/Policy Loss -20.2601 +trainer/Log Pis Mean 23.8918 +trainer/Log Pis Std 13.3186 +trainer/Log Pis Max 61.5825 +trainer/Log Pis Min -9.94744 +trainer/policy/mean Mean -0.0313119 +trainer/policy/mean Std 0.905447 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.84973 +trainer/policy/normal/std Std 0.633708 +trainer/policy/normal/std Max 6.3698 +trainer/policy/normal/std Min 0.332286 +trainer/policy/normal/log_std Mean 1.01238 +trainer/policy/normal/log_std Std 0.296285 +trainer/policy/normal/log_std Max 1.85157 +trainer/policy/normal/log_std Min -1.10176 +eval/num steps total 415385 +eval/num paths total 419 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.225883 +eval/Actions Std 0.945928 +eval/Actions Max 0.999997 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48059 +time/logging (s) 0.00414863 +time/sampling batch (s) 0.530464 +time/saving (s) 0.00417081 +time/training (s) 7.33591 +time/epoch (s) 10.3553 +time/total (s) 4302.7 +Epoch -582 +---------------------------------- --------------- +2022-05-10 14:22:33.309918 PDT | [0] Epoch -581 finished +---------------------------------- --------------- +epoch -581 +replay_buffer/size 999033 +trainer/num train calls 420000 +trainer/Policy Loss -20.123 +trainer/Log Pis Mean 24.5809 +trainer/Log Pis Std 14.657 +trainer/Log Pis Max 75.6728 +trainer/Log Pis Min -9.70513 +trainer/policy/mean Mean -0.0695381 +trainer/policy/mean Std 0.904624 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.83733 +trainer/policy/normal/std Std 0.625889 +trainer/policy/normal/std Max 5.76754 +trainer/policy/normal/std Min 0.343875 +trainer/policy/normal/log_std Mean 1.00766 +trainer/policy/normal/log_std Std 0.299732 +trainer/policy/normal/log_std Max 1.75225 +trainer/policy/normal/log_std Min -1.06748 +eval/num steps total 416385 +eval/num paths total 420 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0220913 +eval/Actions Std 0.911883 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66493 +time/logging (s) 0.00377958 +time/sampling batch (s) 0.279053 +time/saving (s) 0.003704 +time/training (s) 7.09891 +time/epoch (s) 10.0504 +time/total (s) 4312.75 +Epoch -581 +---------------------------------- --------------- +2022-05-10 14:22:43.960629 PDT | [0] Epoch -580 finished +---------------------------------- --------------- +epoch -580 +replay_buffer/size 999033 +trainer/num train calls 421000 +trainer/Policy Loss -20.8596 +trainer/Log Pis Mean 25.2747 +trainer/Log Pis Std 13.2833 +trainer/Log Pis Max 65.0494 +trainer/Log Pis Min -5.42255 +trainer/policy/mean Mean -0.0481848 +trainer/policy/mean Std 0.908882 +trainer/policy/mean Max 0.999973 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85983 +trainer/policy/normal/std Std 0.636886 +trainer/policy/normal/std Max 6.12832 +trainer/policy/normal/std Min 0.307126 +trainer/policy/normal/log_std Mean 1.01545 +trainer/policy/normal/log_std Std 0.298934 +trainer/policy/normal/log_std Max 1.81292 +trainer/policy/normal/log_std Min -1.1805 +eval/num steps total 417385 +eval/num paths total 421 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.129656 +eval/Actions Std 0.839798 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66086 +time/logging (s) 0.00370582 +time/sampling batch (s) 0.528464 +time/saving (s) 0.00343498 +time/training (s) 7.43326 +time/epoch (s) 10.6297 +time/total (s) 4323.39 +Epoch -580 +---------------------------------- --------------- +2022-05-10 14:22:54.360905 PDT | [0] Epoch -579 finished +---------------------------------- --------------- +epoch -579 +replay_buffer/size 999033 +trainer/num train calls 422000 +trainer/Policy Loss -19.3759 +trainer/Log Pis Mean 23.6004 +trainer/Log Pis Std 13.5815 +trainer/Log Pis Max 79.4649 +trainer/Log Pis Min -6.7896 +trainer/policy/mean Mean -0.0279688 +trainer/policy/mean Std 0.903307 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82257 +trainer/policy/normal/std Std 0.622019 +trainer/policy/normal/std Max 5.53756 +trainer/policy/normal/std Min 0.385784 +trainer/policy/normal/log_std Mean 1.00292 +trainer/policy/normal/log_std Std 0.297395 +trainer/policy/normal/log_std Max 1.71155 +trainer/policy/normal/log_std Min -0.952477 +eval/num steps total 418385 +eval/num paths total 422 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.283965 +eval/Actions Std 0.887818 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78014 +time/logging (s) 0.00407923 +time/sampling batch (s) 0.277819 +time/saving (s) 0.0041289 +time/training (s) 7.31372 +time/epoch (s) 10.3799 +time/total (s) 4333.77 +Epoch -579 +---------------------------------- --------------- +2022-05-10 14:23:04.062550 PDT | [0] Epoch -578 finished +---------------------------------- --------------- +epoch -578 +replay_buffer/size 999033 +trainer/num train calls 423000 +trainer/Policy Loss -19.4822 +trainer/Log Pis Mean 23.4827 +trainer/Log Pis Std 13.5959 +trainer/Log Pis Max 68.3892 +trainer/Log Pis Min -6.95609 +trainer/policy/mean Mean -0.0512653 +trainer/policy/mean Std 0.905566 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81852 +trainer/policy/normal/std Std 0.640822 +trainer/policy/normal/std Max 5.49802 +trainer/policy/normal/std Min 0.249755 +trainer/policy/normal/log_std Mean 0.998944 +trainer/policy/normal/log_std Std 0.308625 +trainer/policy/normal/log_std Max 1.70439 +trainer/policy/normal/log_std Min -1.38728 +eval/num steps total 419385 +eval/num paths total 423 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.116508 +eval/Actions Std 0.881911 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64197 +time/logging (s) 0.00365802 +time/sampling batch (s) 0.280727 +time/saving (s) 0.00349547 +time/training (s) 6.75001 +time/epoch (s) 9.67986 +time/total (s) 4343.45 +Epoch -578 +---------------------------------- --------------- +2022-05-10 14:23:13.658727 PDT | [0] Epoch -577 finished +---------------------------------- --------------- +epoch -577 +replay_buffer/size 999033 +trainer/num train calls 424000 +trainer/Policy Loss -19.9777 +trainer/Log Pis Mean 24.2735 +trainer/Log Pis Std 13.111 +trainer/Log Pis Max 71.921 +trainer/Log Pis Min -6.31214 +trainer/policy/mean Mean -0.0631898 +trainer/policy/mean Std 0.904311 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78735 +trainer/policy/normal/std Std 0.635829 +trainer/policy/normal/std Max 5.10875 +trainer/policy/normal/std Min 0.303547 +trainer/policy/normal/log_std Mean 0.986584 +trainer/policy/normal/log_std Std 0.315261 +trainer/policy/normal/log_std Max 1.63096 +trainer/policy/normal/log_std Min -1.19222 +eval/num steps total 420385 +eval/num paths total 424 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.201368 +eval/Actions Std 0.945677 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71362 +time/logging (s) 0.00370713 +time/sampling batch (s) 0.280275 +time/saving (s) 0.00356509 +time/training (s) 6.57405 +time/epoch (s) 9.57521 +time/total (s) 4353.03 +Epoch -577 +---------------------------------- --------------- +2022-05-10 14:23:24.020337 PDT | [0] Epoch -576 finished +---------------------------------- --------------- +epoch -576 +replay_buffer/size 999033 +trainer/num train calls 425000 +trainer/Policy Loss -19.7181 +trainer/Log Pis Mean 23.6207 +trainer/Log Pis Std 13.3656 +trainer/Log Pis Max 68.958 +trainer/Log Pis Min -9.15913 +trainer/policy/mean Mean -0.0409902 +trainer/policy/mean Std 0.904207 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83247 +trainer/policy/normal/std Std 0.604655 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.29458 +trainer/policy/normal/log_std Mean 1.01023 +trainer/policy/normal/log_std Std 0.277305 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22221 +eval/num steps total 421385 +eval/num paths total 425 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.141945 +eval/Actions Std 0.873838 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70967 +time/logging (s) 0.00372203 +time/sampling batch (s) 0.538171 +time/saving (s) 0.00360115 +time/training (s) 7.08506 +time/epoch (s) 10.3402 +time/total (s) 4363.37 +Epoch -576 +---------------------------------- --------------- +2022-05-10 14:23:33.438259 PDT | [0] Epoch -575 finished +---------------------------------- --------------- +epoch -575 +replay_buffer/size 999033 +trainer/num train calls 426000 +trainer/Policy Loss -19.3799 +trainer/Log Pis Mean 24.6452 +trainer/Log Pis Std 13.1418 +trainer/Log Pis Max 73.2852 +trainer/Log Pis Min -8.34006 +trainer/policy/mean Mean -0.0161345 +trainer/policy/mean Std 0.906431 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83703 +trainer/policy/normal/std Std 0.629288 +trainer/policy/normal/std Max 5.35665 +trainer/policy/normal/std Min 0.333696 +trainer/policy/normal/log_std Mean 1.00677 +trainer/policy/normal/log_std Std 0.304269 +trainer/policy/normal/log_std Max 1.67834 +trainer/policy/normal/log_std Min -1.09752 +eval/num steps total 422385 +eval/num paths total 426 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.177861 +eval/Actions Std 0.873777 +eval/Actions Max 0.999998 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61825 +time/logging (s) 0.00416436 +time/sampling batch (s) 0.282339 +time/saving (s) 0.00430572 +time/training (s) 6.48802 +time/epoch (s) 9.39709 +time/total (s) 4372.77 +Epoch -575 +---------------------------------- --------------- +2022-05-10 14:23:43.844148 PDT | [0] Epoch -574 finished +---------------------------------- --------------- +epoch -574 +replay_buffer/size 999033 +trainer/num train calls 427000 +trainer/Policy Loss -19.3878 +trainer/Log Pis Mean 25.0847 +trainer/Log Pis Std 12.9768 +trainer/Log Pis Max 74.5048 +trainer/Log Pis Min -8.11525 +trainer/policy/mean Mean -0.0167551 +trainer/policy/mean Std 0.909143 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82604 +trainer/policy/normal/std Std 0.636949 +trainer/policy/normal/std Max 7.14987 +trainer/policy/normal/std Min 0.317579 +trainer/policy/normal/log_std Mean 1.00324 +trainer/policy/normal/log_std Std 0.298562 +trainer/policy/normal/log_std Max 1.96709 +trainer/policy/normal/log_std Min -1.14703 +eval/num steps total 423385 +eval/num paths total 427 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0487398 +eval/Actions Std 0.867872 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53785 +time/logging (s) 0.00378281 +time/sampling batch (s) 0.531897 +time/saving (s) 0.00356842 +time/training (s) 7.30699 +time/epoch (s) 10.3841 +time/total (s) 4383.16 +Epoch -574 +---------------------------------- --------------- +2022-05-10 14:23:53.328164 PDT | [0] Epoch -573 finished +---------------------------------- --------------- +epoch -573 +replay_buffer/size 999033 +trainer/num train calls 428000 +trainer/Policy Loss -18.3325 +trainer/Log Pis Mean 24.8778 +trainer/Log Pis Std 13.1996 +trainer/Log Pis Max 65.1626 +trainer/Log Pis Min -4.73078 +trainer/policy/mean Mean -0.0364129 +trainer/policy/mean Std 0.90662 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86278 +trainer/policy/normal/std Std 0.637259 +trainer/policy/normal/std Max 6.16844 +trainer/policy/normal/std Min 0.314567 +trainer/policy/normal/log_std Mean 1.01763 +trainer/policy/normal/log_std Std 0.291762 +trainer/policy/normal/log_std Max 1.81945 +trainer/policy/normal/log_std Min -1.15656 +eval/num steps total 424385 +eval/num paths total 428 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102025 +eval/Actions Std 0.918287 +eval/Actions Max 0.999992 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47401 +time/logging (s) 0.00372368 +time/sampling batch (s) 0.277056 +time/saving (s) 0.0034596 +time/training (s) 6.70501 +time/epoch (s) 9.46326 +time/total (s) 4392.63 +Epoch -573 +---------------------------------- --------------- +2022-05-10 14:24:03.760875 PDT | [0] Epoch -572 finished +---------------------------------- --------------- +epoch -572 +replay_buffer/size 999033 +trainer/num train calls 429000 +trainer/Policy Loss -19.6582 +trainer/Log Pis Mean 26.2334 +trainer/Log Pis Std 13.0765 +trainer/Log Pis Max 66.4892 +trainer/Log Pis Min -5.57272 +trainer/policy/mean Mean -0.0593159 +trainer/policy/mean Std 0.905723 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85022 +trainer/policy/normal/std Std 0.637114 +trainer/policy/normal/std Max 5.47628 +trainer/policy/normal/std Min 0.283868 +trainer/policy/normal/log_std Mean 1.0107 +trainer/policy/normal/log_std Std 0.306897 +trainer/policy/normal/log_std Max 1.70043 +trainer/policy/normal/log_std Min -1.25924 +eval/num steps total 425385 +eval/num paths total 429 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.163812 +eval/Actions Std 0.862199 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59911 +time/logging (s) 0.00421041 +time/sampling batch (s) 0.27781 +time/saving (s) 0.00422724 +time/training (s) 7.52696 +time/epoch (s) 10.4123 +time/total (s) 4403.04 +Epoch -572 +---------------------------------- --------------- +2022-05-10 14:24:13.566800 PDT | [0] Epoch -571 finished +---------------------------------- --------------- +epoch -571 +replay_buffer/size 999033 +trainer/num train calls 430000 +trainer/Policy Loss -19.7992 +trainer/Log Pis Mean 25.3364 +trainer/Log Pis Std 12.9437 +trainer/Log Pis Max 76.8856 +trainer/Log Pis Min -10.9258 +trainer/policy/mean Mean -0.0524976 +trainer/policy/mean Std 0.90639 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81787 +trainer/policy/normal/std Std 0.623321 +trainer/policy/normal/std Max 6.03338 +trainer/policy/normal/std Min 0.33296 +trainer/policy/normal/log_std Mean 1.00055 +trainer/policy/normal/log_std Std 0.30074 +trainer/policy/normal/log_std Max 1.79731 +trainer/policy/normal/log_std Min -1.09973 +eval/num steps total 426385 +eval/num paths total 430 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.184221 +eval/Actions Std 0.905882 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70787 +time/logging (s) 0.00380133 +time/sampling batch (s) 0.276905 +time/saving (s) 0.00362113 +time/training (s) 6.79205 +time/epoch (s) 9.78425 +time/total (s) 4412.83 +Epoch -571 +---------------------------------- --------------- +2022-05-10 14:24:23.114011 PDT | [0] Epoch -570 finished +---------------------------------- --------------- +epoch -570 +replay_buffer/size 999033 +trainer/num train calls 431000 +trainer/Policy Loss -19.2229 +trainer/Log Pis Mean 24.6907 +trainer/Log Pis Std 12.995 +trainer/Log Pis Max 69.3528 +trainer/Log Pis Min -6.70563 +trainer/policy/mean Mean -0.0302279 +trainer/policy/mean Std 0.909618 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.7949 +trainer/policy/normal/std Std 0.63953 +trainer/policy/normal/std Max 6.21409 +trainer/policy/normal/std Min 0.238385 +trainer/policy/normal/log_std Mean 0.989773 +trainer/policy/normal/log_std Std 0.312215 +trainer/policy/normal/log_std Max 1.82682 +trainer/policy/normal/log_std Min -1.43387 +eval/num steps total 427385 +eval/num paths total 431 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0254711 +eval/Actions Std 0.907494 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34552 +time/logging (s) 0.00369571 +time/sampling batch (s) 0.527074 +time/saving (s) 0.00343828 +time/training (s) 6.64664 +time/epoch (s) 9.52637 +time/total (s) 4422.36 +Epoch -570 +---------------------------------- --------------- +2022-05-10 14:24:34.126959 PDT | [0] Epoch -569 finished +---------------------------------- --------------- +epoch -569 +replay_buffer/size 999033 +trainer/num train calls 432000 +trainer/Policy Loss -20.0947 +trainer/Log Pis Mean 24.5506 +trainer/Log Pis Std 13.217 +trainer/Log Pis Max 68.5505 +trainer/Log Pis Min -12.58 +trainer/policy/mean Mean -0.0554929 +trainer/policy/mean Std 0.90941 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.82883 +trainer/policy/normal/std Std 0.625665 +trainer/policy/normal/std Max 6.54328 +trainer/policy/normal/std Min 0.286757 +trainer/policy/normal/log_std Mean 1.00514 +trainer/policy/normal/log_std Std 0.296696 +trainer/policy/normal/log_std Max 1.87844 +trainer/policy/normal/log_std Min -1.24912 +eval/num steps total 428385 +eval/num paths total 432 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179752 +eval/Actions Std 0.896895 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66974 +time/logging (s) 0.00415839 +time/sampling batch (s) 0.278771 +time/saving (s) 0.00408269 +time/training (s) 8.03579 +time/epoch (s) 10.9925 +time/total (s) 4433.35 +Epoch -569 +---------------------------------- --------------- +2022-05-10 14:24:44.734787 PDT | [0] Epoch -568 finished +---------------------------------- --------------- +epoch -568 +replay_buffer/size 999033 +trainer/num train calls 433000 +trainer/Policy Loss -19.9149 +trainer/Log Pis Mean 24.5343 +trainer/Log Pis Std 13.1763 +trainer/Log Pis Max 72.7101 +trainer/Log Pis Min -11.6104 +trainer/policy/mean Mean -0.0264199 +trainer/policy/mean Std 0.909108 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.77652 +trainer/policy/normal/std Std 0.62715 +trainer/policy/normal/std Max 5.68836 +trainer/policy/normal/std Min 0.291522 +trainer/policy/normal/log_std Mean 0.983547 +trainer/policy/normal/log_std Std 0.311729 +trainer/policy/normal/log_std Max 1.73842 +trainer/policy/normal/log_std Min -1.23264 +eval/num steps total 429385 +eval/num paths total 433 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.221809 +eval/Actions Std 0.909561 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73776 +time/logging (s) 0.00377661 +time/sampling batch (s) 0.279307 +time/saving (s) 0.00366426 +time/training (s) 7.56145 +time/epoch (s) 10.586 +time/total (s) 4443.94 +Epoch -568 +---------------------------------- --------------- +2022-05-10 14:24:55.239238 PDT | [0] Epoch -567 finished +---------------------------------- --------------- +epoch -567 +replay_buffer/size 999033 +trainer/num train calls 434000 +trainer/Policy Loss -19.9112 +trainer/Log Pis Mean 23.6362 +trainer/Log Pis Std 12.9406 +trainer/Log Pis Max 72.5343 +trainer/Log Pis Min -11.6023 +trainer/policy/mean Mean -0.0292072 +trainer/policy/mean Std 0.904108 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.77125 +trainer/policy/normal/std Std 0.627182 +trainer/policy/normal/std Max 5.28556 +trainer/policy/normal/std Min 0.291164 +trainer/policy/normal/log_std Mean 0.982162 +trainer/policy/normal/log_std Std 0.307891 +trainer/policy/normal/log_std Max 1.66498 +trainer/policy/normal/log_std Min -1.23387 +eval/num steps total 430385 +eval/num paths total 434 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.157265 +eval/Actions Std 0.933789 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66531 +time/logging (s) 0.00364818 +time/sampling batch (s) 0.285999 +time/saving (s) 0.00346664 +time/training (s) 7.52445 +time/epoch (s) 10.4829 +time/total (s) 4454.43 +Epoch -567 +---------------------------------- --------------- +2022-05-10 14:25:06.615695 PDT | [0] Epoch -566 finished +---------------------------------- --------------- +epoch -566 +replay_buffer/size 999033 +trainer/num train calls 435000 +trainer/Policy Loss -19.3024 +trainer/Log Pis Mean 25.1152 +trainer/Log Pis Std 12.7682 +trainer/Log Pis Max 63.052 +trainer/Log Pis Min -5.7876 +trainer/policy/mean Mean -0.021404 +trainer/policy/mean Std 0.908417 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75556 +trainer/policy/normal/std Std 0.63369 +trainer/policy/normal/std Max 5.55039 +trainer/policy/normal/std Min 0.320919 +trainer/policy/normal/log_std Mean 0.9739 +trainer/policy/normal/log_std Std 0.321551 +trainer/policy/normal/log_std Max 1.71387 +trainer/policy/normal/log_std Min -1.13657 +eval/num steps total 431385 +eval/num paths total 435 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.156501 +eval/Actions Std 0.869582 +eval/Actions Max 0.99999 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58687 +time/logging (s) 0.00417615 +time/sampling batch (s) 0.283068 +time/saving (s) 0.00432074 +time/training (s) 8.47726 +time/epoch (s) 11.3557 +time/total (s) 4465.79 +Epoch -566 +---------------------------------- --------------- +2022-05-10 14:25:16.401166 PDT | [0] Epoch -565 finished +---------------------------------- --------------- +epoch -565 +replay_buffer/size 999033 +trainer/num train calls 436000 +trainer/Policy Loss -20.4282 +trainer/Log Pis Mean 24.3926 +trainer/Log Pis Std 13.3165 +trainer/Log Pis Max 72.796 +trainer/Log Pis Min -6.61281 +trainer/policy/mean Mean -0.0365339 +trainer/policy/mean Std 0.910631 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78385 +trainer/policy/normal/std Std 0.630923 +trainer/policy/normal/std Max 5.64874 +trainer/policy/normal/std Min 0.306571 +trainer/policy/normal/log_std Mean 0.986304 +trainer/policy/normal/log_std Std 0.310325 +trainer/policy/normal/log_std Max 1.73143 +trainer/policy/normal/log_std Min -1.18231 +eval/num steps total 432385 +eval/num paths total 436 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00323433 +eval/Actions Std 0.911012 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60088 +time/logging (s) 0.00375069 +time/sampling batch (s) 0.279604 +time/saving (s) 0.00361011 +time/training (s) 6.87566 +time/epoch (s) 9.7635 +time/total (s) 4475.55 +Epoch -565 +---------------------------------- --------------- +2022-05-10 14:25:26.817723 PDT | [0] Epoch -564 finished +---------------------------------- --------------- +epoch -564 +replay_buffer/size 999033 +trainer/num train calls 437000 +trainer/Policy Loss -18.8548 +trainer/Log Pis Mean 24.3214 +trainer/Log Pis Std 13.0662 +trainer/Log Pis Max 71.625 +trainer/Log Pis Min -6.1781 +trainer/policy/mean Mean -0.0461969 +trainer/policy/mean Std 0.90323 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76581 +trainer/policy/normal/std Std 0.600209 +trainer/policy/normal/std Max 5.43414 +trainer/policy/normal/std Min 0.253994 +trainer/policy/normal/log_std Mean 0.983978 +trainer/policy/normal/log_std Std 0.289701 +trainer/policy/normal/log_std Max 1.6927 +trainer/policy/normal/log_std Min -1.37044 +eval/num steps total 433385 +eval/num paths total 437 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0143781 +eval/Actions Std 0.905852 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63851 +time/logging (s) 0.0037912 +time/sampling batch (s) 0.282326 +time/saving (s) 0.00345521 +time/training (s) 7.46727 +time/epoch (s) 10.3953 +time/total (s) 4485.95 +Epoch -564 +---------------------------------- --------------- +2022-05-10 14:25:36.871200 PDT | [0] Epoch -563 finished +---------------------------------- --------------- +epoch -563 +replay_buffer/size 999033 +trainer/num train calls 438000 +trainer/Policy Loss -18.7382 +trainer/Log Pis Mean 24.2067 +trainer/Log Pis Std 13.3541 +trainer/Log Pis Max 62.1093 +trainer/Log Pis Min -9.84737 +trainer/policy/mean Mean -0.00607009 +trainer/policy/mean Std 0.905095 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75907 +trainer/policy/normal/std Std 0.629677 +trainer/policy/normal/std Max 5.44674 +trainer/policy/normal/std Min 0.285856 +trainer/policy/normal/log_std Mean 0.978244 +trainer/policy/normal/log_std Std 0.303152 +trainer/policy/normal/log_std Max 1.69502 +trainer/policy/normal/log_std Min -1.25227 +eval/num steps total 434385 +eval/num paths total 438 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.183569 +eval/Actions Std 0.922931 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5069 +time/logging (s) 0.00384178 +time/sampling batch (s) 0.535128 +time/saving (s) 0.0035523 +time/training (s) 6.98268 +time/epoch (s) 10.0321 +time/total (s) 4495.99 +Epoch -563 +---------------------------------- --------------- +2022-05-10 14:25:46.517486 PDT | [0] Epoch -562 finished +---------------------------------- --------------- +epoch -562 +replay_buffer/size 999033 +trainer/num train calls 439000 +trainer/Policy Loss -19.8112 +trainer/Log Pis Mean 24.9153 +trainer/Log Pis Std 13.246 +trainer/Log Pis Max 72.2324 +trainer/Log Pis Min -8.21072 +trainer/policy/mean Mean -0.0482955 +trainer/policy/mean Std 0.910754 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.84365 +trainer/policy/normal/std Std 0.643718 +trainer/policy/normal/std Max 6.64747 +trainer/policy/normal/std Min 0.325159 +trainer/policy/normal/log_std Mean 1.00967 +trainer/policy/normal/log_std Std 0.297027 +trainer/policy/normal/log_std Max 1.89424 +trainer/policy/normal/log_std Min -1.12344 +eval/num steps total 435385 +eval/num paths total 439 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.061296 +eval/Actions Std 0.902031 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71784 +time/logging (s) 0.0041761 +time/sampling batch (s) 0.279959 +time/saving (s) 0.00431253 +time/training (s) 6.61919 +time/epoch (s) 9.62548 +time/total (s) 4505.62 +Epoch -562 +---------------------------------- --------------- +2022-05-10 14:25:56.422370 PDT | [0] Epoch -561 finished +---------------------------------- --------------- +epoch -561 +replay_buffer/size 999033 +trainer/num train calls 440000 +trainer/Policy Loss -21.493 +trainer/Log Pis Mean 23.8191 +trainer/Log Pis Std 13.2296 +trainer/Log Pis Max 59.2329 +trainer/Log Pis Min -10.1439 +trainer/policy/mean Mean -0.0454525 +trainer/policy/mean Std 0.907783 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80794 +trainer/policy/normal/std Std 0.612225 +trainer/policy/normal/std Max 5.24081 +trainer/policy/normal/std Min 0.343558 +trainer/policy/normal/log_std Mean 0.997574 +trainer/policy/normal/log_std Std 0.299138 +trainer/policy/normal/log_std Max 1.65648 +trainer/policy/normal/log_std Min -1.0684 +eval/num steps total 436385 +eval/num paths total 440 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.12598 +eval/Actions Std 0.907238 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63271 +time/logging (s) 0.00366638 +time/sampling batch (s) 0.27799 +time/saving (s) 0.00347439 +time/training (s) 6.96521 +time/epoch (s) 9.88305 +time/total (s) 4515.5 +Epoch -561 +---------------------------------- --------------- +2022-05-10 14:26:08.191362 PDT | [0] Epoch -560 finished +---------------------------------- --------------- +epoch -560 +replay_buffer/size 999033 +trainer/num train calls 441000 +trainer/Policy Loss -19.2062 +trainer/Log Pis Mean 24.1468 +trainer/Log Pis Std 12.7106 +trainer/Log Pis Max 67.1416 +trainer/Log Pis Min -9.55386 +trainer/policy/mean Mean -0.0292646 +trainer/policy/mean Std 0.904121 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.78761 +trainer/policy/normal/std Std 0.641433 +trainer/policy/normal/std Max 5.03068 +trainer/policy/normal/std Min 0.356332 +trainer/policy/normal/log_std Mean 0.986725 +trainer/policy/normal/log_std Std 0.31273 +trainer/policy/normal/log_std Max 1.61556 +trainer/policy/normal/log_std Min -1.03189 +eval/num steps total 437385 +eval/num paths total 441 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116278 +eval/Actions Std 0.910201 +eval/Actions Max 0.999988 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55851 +time/logging (s) 0.00367394 +time/sampling batch (s) 0.53379 +time/saving (s) 0.00344955 +time/training (s) 8.64845 +time/epoch (s) 11.7479 +time/total (s) 4527.25 +Epoch -560 +---------------------------------- --------------- +2022-05-10 14:26:18.294934 PDT | [0] Epoch -559 finished +---------------------------------- --------------- +epoch -559 +replay_buffer/size 999033 +trainer/num train calls 442000 +trainer/Policy Loss -18.9194 +trainer/Log Pis Mean 23.1901 +trainer/Log Pis Std 13.3576 +trainer/Log Pis Max 70.1044 +trainer/Log Pis Min -6.45087 +trainer/policy/mean Mean -0.0337144 +trainer/policy/mean Std 0.906746 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.84501 +trainer/policy/normal/std Std 0.648721 +trainer/policy/normal/std Max 6.04892 +trainer/policy/normal/std Min 0.303 +trainer/policy/normal/log_std Mean 1.00926 +trainer/policy/normal/log_std Std 0.30032 +trainer/policy/normal/log_std Max 1.79988 +trainer/policy/normal/log_std Min -1.19402 +eval/num steps total 438385 +eval/num paths total 442 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.219586 +eval/Actions Std 0.895248 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59256 +time/logging (s) 0.00434163 +time/sampling batch (s) 0.530672 +time/saving (s) 0.00439935 +time/training (s) 6.95109 +time/epoch (s) 10.0831 +time/total (s) 4537.34 +Epoch -559 +---------------------------------- --------------- +2022-05-10 14:26:28.621798 PDT | [0] Epoch -558 finished +---------------------------------- --------------- +epoch -558 +replay_buffer/size 999033 +trainer/num train calls 443000 +trainer/Policy Loss -19.8069 +trainer/Log Pis Mean 24.2167 +trainer/Log Pis Std 13.2599 +trainer/Log Pis Max 76.6997 +trainer/Log Pis Min -12.8818 +trainer/policy/mean Mean -0.057895 +trainer/policy/mean Std 0.907055 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.81363 +trainer/policy/normal/std Std 0.645796 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.295226 +trainer/policy/normal/log_std Mean 0.995434 +trainer/policy/normal/log_std Std 0.318887 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22001 +eval/num steps total 439385 +eval/num paths total 443 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0485238 +eval/Actions Std 0.907818 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63892 +time/logging (s) 0.00376232 +time/sampling batch (s) 0.283406 +time/saving (s) 0.00395569 +time/training (s) 7.37433 +time/epoch (s) 10.3044 +time/total (s) 4547.65 +Epoch -558 +---------------------------------- --------------- +2022-05-10 14:26:38.391298 PDT | [0] Epoch -557 finished +---------------------------------- --------------- +epoch -557 +replay_buffer/size 999033 +trainer/num train calls 444000 +trainer/Policy Loss -19.3211 +trainer/Log Pis Mean 24.7967 +trainer/Log Pis Std 13.9964 +trainer/Log Pis Max 71.3881 +trainer/Log Pis Min -8.35869 +trainer/policy/mean Mean -0.0522935 +trainer/policy/mean Std 0.906824 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.86993 +trainer/policy/normal/std Std 0.619207 +trainer/policy/normal/std Max 6.5464 +trainer/policy/normal/std Min 0.345455 +trainer/policy/normal/log_std Mean 1.02212 +trainer/policy/normal/log_std Std 0.281657 +trainer/policy/normal/log_std Max 1.87892 +trainer/policy/normal/log_std Min -1.06289 +eval/num steps total 440385 +eval/num paths total 444 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0590184 +eval/Actions Std 0.907244 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57086 +time/logging (s) 0.00368444 +time/sampling batch (s) 0.281857 +time/saving (s) 0.00346762 +time/training (s) 6.88841 +time/epoch (s) 9.74828 +time/total (s) 4557.4 +Epoch -557 +---------------------------------- --------------- +2022-05-10 14:26:48.705460 PDT | [0] Epoch -556 finished +---------------------------------- --------------- +epoch -556 +replay_buffer/size 999033 +trainer/num train calls 445000 +trainer/Policy Loss -19.0969 +trainer/Log Pis Mean 25.2922 +trainer/Log Pis Std 13.5266 +trainer/Log Pis Max 80.6341 +trainer/Log Pis Min -3.23474 +trainer/policy/mean Mean -0.0429985 +trainer/policy/mean Std 0.904 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81499 +trainer/policy/normal/std Std 0.628648 +trainer/policy/normal/std Max 5.84092 +trainer/policy/normal/std Min 0.30834 +trainer/policy/normal/log_std Mean 0.99892 +trainer/policy/normal/log_std Std 0.302956 +trainer/policy/normal/log_std Max 1.76489 +trainer/policy/normal/log_std Min -1.17655 +eval/num steps total 441385 +eval/num paths total 445 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0010403 +eval/Actions Std 0.911103 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6523 +time/logging (s) 0.00404653 +time/sampling batch (s) 0.283153 +time/saving (s) 0.00374788 +time/training (s) 7.34989 +time/epoch (s) 10.2931 +time/total (s) 4567.69 +Epoch -556 +---------------------------------- --------------- +2022-05-10 14:26:59.304706 PDT | [0] Epoch -555 finished +---------------------------------- --------------- +epoch -555 +replay_buffer/size 999033 +trainer/num train calls 446000 +trainer/Policy Loss -19.0372 +trainer/Log Pis Mean 25.3869 +trainer/Log Pis Std 12.9225 +trainer/Log Pis Max 65.578 +trainer/Log Pis Min -7.71164 +trainer/policy/mean Mean -0.0460172 +trainer/policy/mean Std 0.909754 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82791 +trainer/policy/normal/std Std 0.621941 +trainer/policy/normal/std Max 6.06456 +trainer/policy/normal/std Min 0.314415 +trainer/policy/normal/log_std Mean 1.00461 +trainer/policy/normal/log_std Std 0.297516 +trainer/policy/normal/log_std Max 1.80246 +trainer/policy/normal/log_std Min -1.15704 +eval/num steps total 442385 +eval/num paths total 446 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.303517 +eval/Actions Std 0.869444 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69193 +time/logging (s) 0.00394589 +time/sampling batch (s) 0.283457 +time/saving (s) 0.0041581 +time/training (s) 7.59422 +time/epoch (s) 10.5777 +time/total (s) 4578.27 +Epoch -555 +---------------------------------- --------------- +2022-05-10 14:27:09.813918 PDT | [0] Epoch -554 finished +---------------------------------- --------------- +epoch -554 +replay_buffer/size 999033 +trainer/num train calls 447000 +trainer/Policy Loss -19.7971 +trainer/Log Pis Mean 23.8177 +trainer/Log Pis Std 13.2131 +trainer/Log Pis Max 69.3683 +trainer/Log Pis Min -8.19919 +trainer/policy/mean Mean -0.035351 +trainer/policy/mean Std 0.906728 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7647 +trainer/policy/normal/std Std 0.632966 +trainer/policy/normal/std Max 6.37768 +trainer/policy/normal/std Min 0.324257 +trainer/policy/normal/log_std Mean 0.978813 +trainer/policy/normal/log_std Std 0.311041 +trainer/policy/normal/log_std Max 1.8528 +trainer/policy/normal/log_std Min -1.12622 +eval/num steps total 443385 +eval/num paths total 447 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0235673 +eval/Actions Std 0.904454 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64271 +time/logging (s) 0.00381233 +time/sampling batch (s) 0.283843 +time/saving (s) 0.00350746 +time/training (s) 7.55328 +time/epoch (s) 10.4871 +time/total (s) 4588.76 +Epoch -554 +---------------------------------- --------------- +2022-05-10 14:27:20.279606 PDT | [0] Epoch -553 finished +---------------------------------- --------------- +epoch -553 +replay_buffer/size 999033 +trainer/num train calls 448000 +trainer/Policy Loss -20.4319 +trainer/Log Pis Mean 24.1532 +trainer/Log Pis Std 12.8875 +trainer/Log Pis Max 69.8122 +trainer/Log Pis Min -4.70156 +trainer/policy/mean Mean -0.0284642 +trainer/policy/mean Std 0.905706 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.8308 +trainer/policy/normal/std Std 0.653056 +trainer/policy/normal/std Max 6.284 +trainer/policy/normal/std Min 0.28784 +trainer/policy/normal/log_std Mean 1.00185 +trainer/policy/normal/log_std Std 0.315549 +trainer/policy/normal/log_std Max 1.83801 +trainer/policy/normal/log_std Min -1.24535 +eval/num steps total 444385 +eval/num paths total 448 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.274752 +eval/Actions Std 0.915776 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63055 +time/logging (s) 0.00372354 +time/sampling batch (s) 0.283336 +time/saving (s) 0.00346113 +time/training (s) 7.52287 +time/epoch (s) 10.4439 +time/total (s) 4599.21 +Epoch -553 +---------------------------------- --------------- +2022-05-10 14:27:30.832582 PDT | [0] Epoch -552 finished +---------------------------------- --------------- +epoch -552 +replay_buffer/size 999033 +trainer/num train calls 449000 +trainer/Policy Loss -19.7198 +trainer/Log Pis Mean 24.9281 +trainer/Log Pis Std 13.7366 +trainer/Log Pis Max 76.3802 +trainer/Log Pis Min -12.9072 +trainer/policy/mean Mean -0.0260992 +trainer/policy/mean Std 0.906469 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81642 +trainer/policy/normal/std Std 0.639098 +trainer/policy/normal/std Max 5.96392 +trainer/policy/normal/std Min 0.300931 +trainer/policy/normal/log_std Mean 0.999004 +trainer/policy/normal/log_std Std 0.303085 +trainer/policy/normal/log_std Max 1.78573 +trainer/policy/normal/log_std Min -1.20087 +eval/num steps total 445385 +eval/num paths total 449 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.214687 +eval/Actions Std 0.941257 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61706 +time/logging (s) 0.00431006 +time/sampling batch (s) 0.536198 +time/saving (s) 0.00417935 +time/training (s) 7.37028 +time/epoch (s) 10.532 +time/total (s) 4609.75 +Epoch -552 +---------------------------------- --------------- +2022-05-10 14:27:40.885439 PDT | [0] Epoch -551 finished +---------------------------------- --------------- +epoch -551 +replay_buffer/size 999033 +trainer/num train calls 450000 +trainer/Policy Loss -19.1679 +trainer/Log Pis Mean 24.6351 +trainer/Log Pis Std 13.4497 +trainer/Log Pis Max 80.4311 +trainer/Log Pis Min -11.5099 +trainer/policy/mean Mean -0.0303668 +trainer/policy/mean Std 0.903972 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85183 +trainer/policy/normal/std Std 0.662924 +trainer/policy/normal/std Max 5.38263 +trainer/policy/normal/std Min 0.253299 +trainer/policy/normal/log_std Mean 1.00625 +trainer/policy/normal/log_std Std 0.333239 +trainer/policy/normal/log_std Max 1.68318 +trainer/policy/normal/log_std Min -1.37318 +eval/num steps total 446385 +eval/num paths total 450 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0999405 +eval/Actions Std 0.889095 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55394 +time/logging (s) 0.00376094 +time/sampling batch (s) 0.283074 +time/saving (s) 0.00366124 +time/training (s) 7.18583 +time/epoch (s) 10.0303 +time/total (s) 4619.78 +Epoch -551 +---------------------------------- --------------- +2022-05-10 14:27:50.528653 PDT | [0] Epoch -550 finished +---------------------------------- --------------- +epoch -550 +replay_buffer/size 999033 +trainer/num train calls 451000 +trainer/Policy Loss -20.1813 +trainer/Log Pis Mean 24.7356 +trainer/Log Pis Std 13.722 +trainer/Log Pis Max 72.1528 +trainer/Log Pis Min -13.2102 +trainer/policy/mean Mean -0.0211742 +trainer/policy/mean Std 0.903957 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85864 +trainer/policy/normal/std Std 0.640995 +trainer/policy/normal/std Max 6.12608 +trainer/policy/normal/std Min 0.261702 +trainer/policy/normal/log_std Mean 1.0143 +trainer/policy/normal/log_std Std 0.303803 +trainer/policy/normal/log_std Max 1.81255 +trainer/policy/normal/log_std Min -1.34055 +eval/num steps total 447385 +eval/num paths total 451 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.207372 +eval/Actions Std 0.894182 +eval/Actions Max 1 +eval/Actions Min -0.999956 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61359 +time/logging (s) 0.00371645 +time/sampling batch (s) 0.277517 +time/saving (s) 0.00350054 +time/training (s) 6.7238 +time/epoch (s) 9.62213 +time/total (s) 4629.4 +Epoch -550 +---------------------------------- --------------- +2022-05-10 14:28:01.025396 PDT | [0] Epoch -549 finished +---------------------------------- --------------- +epoch -549 +replay_buffer/size 999033 +trainer/num train calls 452000 +trainer/Policy Loss -18.8177 +trainer/Log Pis Mean 24.5189 +trainer/Log Pis Std 13.4016 +trainer/Log Pis Max 79.3854 +trainer/Log Pis Min -11.4899 +trainer/policy/mean Mean -0.0357712 +trainer/policy/mean Std 0.902204 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.88622 +trainer/policy/normal/std Std 0.669471 +trainer/policy/normal/std Max 6.38964 +trainer/policy/normal/std Min 0.29432 +trainer/policy/normal/log_std Mean 1.02058 +trainer/policy/normal/log_std Std 0.317717 +trainer/policy/normal/log_std Max 1.85468 +trainer/policy/normal/log_std Min -1.22309 +eval/num steps total 448385 +eval/num paths total 452 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.357663 +eval/Actions Std 0.681691 +eval/Actions Max 0.999989 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50312 +time/logging (s) 0.00379103 +time/sampling batch (s) 0.278585 +time/saving (s) 0.00352142 +time/training (s) 7.68685 +time/epoch (s) 10.4759 +time/total (s) 4639.88 +Epoch -549 +---------------------------------- --------------- +2022-05-10 14:28:11.007935 PDT | [0] Epoch -548 finished +---------------------------------- --------------- +epoch -548 +replay_buffer/size 999033 +trainer/num train calls 453000 +trainer/Policy Loss -17.9551 +trainer/Log Pis Mean 24.1967 +trainer/Log Pis Std 13.9489 +trainer/Log Pis Max 64.1972 +trainer/Log Pis Min -7.48108 +trainer/policy/mean Mean -0.0455944 +trainer/policy/mean Std 0.906348 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84004 +trainer/policy/normal/std Std 0.628845 +trainer/policy/normal/std Max 5.17389 +trainer/policy/normal/std Min 0.300133 +trainer/policy/normal/log_std Mean 1.00861 +trainer/policy/normal/log_std Std 0.299706 +trainer/policy/normal/log_std Max 1.64363 +trainer/policy/normal/log_std Min -1.20353 +eval/num steps total 449385 +eval/num paths total 453 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0155502 +eval/Actions Std 0.905269 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62718 +time/logging (s) 0.00370533 +time/sampling batch (s) 0.27818 +time/saving (s) 0.00348803 +time/training (s) 7.04898 +time/epoch (s) 9.96152 +time/total (s) 4649.85 +Epoch -548 +---------------------------------- --------------- +2022-05-10 14:28:20.679880 PDT | [0] Epoch -547 finished +---------------------------------- --------------- +epoch -547 +replay_buffer/size 999033 +trainer/num train calls 454000 +trainer/Policy Loss -19.4122 +trainer/Log Pis Mean 23.9175 +trainer/Log Pis Std 13.1369 +trainer/Log Pis Max 65.6597 +trainer/Log Pis Min -9.61642 +trainer/policy/mean Mean -0.0281262 +trainer/policy/mean Std 0.906954 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80477 +trainer/policy/normal/std Std 0.637759 +trainer/policy/normal/std Max 5.29844 +trainer/policy/normal/std Min 0.358621 +trainer/policy/normal/log_std Mean 0.9941 +trainer/policy/normal/log_std Std 0.307405 +trainer/policy/normal/log_std Max 1.66741 +trainer/policy/normal/log_std Min -1.02549 +eval/num steps total 450385 +eval/num paths total 454 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109324 +eval/Actions Std 0.912094 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59334 +time/logging (s) 0.00368655 +time/sampling batch (s) 0.279291 +time/saving (s) 0.00343571 +time/training (s) 6.77118 +time/epoch (s) 9.65094 +time/total (s) 4659.5 +Epoch -547 +---------------------------------- --------------- +2022-05-10 14:28:30.686445 PDT | [0] Epoch -546 finished +---------------------------------- --------------- +epoch -546 +replay_buffer/size 999033 +trainer/num train calls 455000 +trainer/Policy Loss -19.3939 +trainer/Log Pis Mean 24.5632 +trainer/Log Pis Std 13.123 +trainer/Log Pis Max 75.9382 +trainer/Log Pis Min -8.94319 +trainer/policy/mean Mean -0.0374346 +trainer/policy/mean Std 0.905201 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7771 +trainer/policy/normal/std Std 0.604481 +trainer/policy/normal/std Max 5.3611 +trainer/policy/normal/std Min 0.315206 +trainer/policy/normal/log_std Mean 0.98762 +trainer/policy/normal/log_std Std 0.292397 +trainer/policy/normal/log_std Max 1.67917 +trainer/policy/normal/log_std Min -1.15453 +eval/num steps total 451385 +eval/num paths total 455 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0946765 +eval/Actions Std 0.893785 +eval/Actions Max 0.999998 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61214 +time/logging (s) 0.00413549 +time/sampling batch (s) 0.533517 +time/saving (s) 0.00410919 +time/training (s) 6.83169 +time/epoch (s) 9.98559 +time/total (s) 4669.49 +Epoch -546 +---------------------------------- --------------- +2022-05-10 14:28:40.204861 PDT | [0] Epoch -545 finished +---------------------------------- --------------- +epoch -545 +replay_buffer/size 999033 +trainer/num train calls 456000 +trainer/Policy Loss -18.8386 +trainer/Log Pis Mean 24.5825 +trainer/Log Pis Std 12.9022 +trainer/Log Pis Max 65.7205 +trainer/Log Pis Min -4.50186 +trainer/policy/mean Mean -0.0323019 +trainer/policy/mean Std 0.907118 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.85807 +trainer/policy/normal/std Std 0.633725 +trainer/policy/normal/std Max 5.81767 +trainer/policy/normal/std Min 0.277915 +trainer/policy/normal/log_std Mean 1.01533 +trainer/policy/normal/log_std Std 0.296651 +trainer/policy/normal/log_std Max 1.7609 +trainer/policy/normal/log_std Min -1.28044 +eval/num steps total 452385 +eval/num paths total 456 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0181534 +eval/Actions Std 0.918681 +eval/Actions Max 0.999991 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66475 +time/logging (s) 0.00385919 +time/sampling batch (s) 0.52919 +time/saving (s) 0.00368599 +time/training (s) 6.29535 +time/epoch (s) 9.49683 +time/total (s) 4678.99 +Epoch -545 +---------------------------------- --------------- +2022-05-10 14:28:50.632354 PDT | [0] Epoch -544 finished +---------------------------------- --------------- +epoch -544 +replay_buffer/size 999033 +trainer/num train calls 457000 +trainer/Policy Loss -18.7284 +trainer/Log Pis Mean 24.2747 +trainer/Log Pis Std 13.0182 +trainer/Log Pis Max 73.6945 +trainer/Log Pis Min -6.30339 +trainer/policy/mean Mean -0.0501345 +trainer/policy/mean Std 0.907424 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79769 +trainer/policy/normal/std Std 0.629634 +trainer/policy/normal/std Max 5.03914 +trainer/policy/normal/std Min 0.274331 +trainer/policy/normal/log_std Mean 0.991051 +trainer/policy/normal/log_std Std 0.313147 +trainer/policy/normal/log_std Max 1.61723 +trainer/policy/normal/log_std Min -1.29342 +eval/num steps total 453385 +eval/num paths total 457 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.127508 +eval/Actions Std 0.920039 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60744 +time/logging (s) 0.00375819 +time/sampling batch (s) 0.529584 +time/saving (s) 0.00349847 +time/training (s) 7.26189 +time/epoch (s) 10.4062 +time/total (s) 4689.4 +Epoch -544 +---------------------------------- --------------- +2022-05-10 14:29:00.235633 PDT | [0] Epoch -543 finished +---------------------------------- --------------- +epoch -543 +replay_buffer/size 999033 +trainer/num train calls 458000 +trainer/Policy Loss -19.3582 +trainer/Log Pis Mean 23.9307 +trainer/Log Pis Std 13.2036 +trainer/Log Pis Max 68.717 +trainer/Log Pis Min -9.69304 +trainer/policy/mean Mean -0.00889915 +trainer/policy/mean Std 0.907236 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78881 +trainer/policy/normal/std Std 0.634377 +trainer/policy/normal/std Max 4.86816 +trainer/policy/normal/std Min 0.355768 +trainer/policy/normal/log_std Mean 0.988208 +trainer/policy/normal/log_std Std 0.307693 +trainer/policy/normal/log_std Max 1.58272 +trainer/policy/normal/log_std Min -1.03348 +eval/num steps total 454385 +eval/num paths total 458 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.358598 +eval/Actions Std 0.842062 +eval/Actions Max 1 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54023 +time/logging (s) 0.00390693 +time/sampling batch (s) 0.530187 +time/saving (s) 0.00347407 +time/training (s) 6.50447 +time/epoch (s) 9.58227 +time/total (s) 4698.98 +Epoch -543 +---------------------------------- --------------- +2022-05-10 14:29:11.038226 PDT | [0] Epoch -542 finished +---------------------------------- --------------- +epoch -542 +replay_buffer/size 999033 +trainer/num train calls 459000 +trainer/Policy Loss -20.4123 +trainer/Log Pis Mean 24.2664 +trainer/Log Pis Std 13.0177 +trainer/Log Pis Max 69.1166 +trainer/Log Pis Min -6.40332 +trainer/policy/mean Mean -0.0294786 +trainer/policy/mean Std 0.903367 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.83287 +trainer/policy/normal/std Std 0.649816 +trainer/policy/normal/std Max 5.70929 +trainer/policy/normal/std Min 0.252391 +trainer/policy/normal/log_std Mean 1.00237 +trainer/policy/normal/log_std Std 0.317795 +trainer/policy/normal/log_std Max 1.7421 +trainer/policy/normal/log_std Min -1.37677 +eval/num steps total 455385 +eval/num paths total 459 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128749 +eval/Actions Std 0.898088 +eval/Actions Max 0.999985 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79835 +time/logging (s) 0.00374584 +time/sampling batch (s) 0.281619 +time/saving (s) 0.00345288 +time/training (s) 7.69405 +time/epoch (s) 10.7812 +time/total (s) 4709.77 +Epoch -542 +---------------------------------- --------------- +2022-05-10 14:29:20.074460 PDT | [0] Epoch -541 finished +---------------------------------- --------------- +epoch -541 +replay_buffer/size 999033 +trainer/num train calls 460000 +trainer/Policy Loss -19.5186 +trainer/Log Pis Mean 24.4622 +trainer/Log Pis Std 13.2619 +trainer/Log Pis Max 72.7673 +trainer/Log Pis Min -8.56606 +trainer/policy/mean Mean -0.0315384 +trainer/policy/mean Std 0.906923 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86922 +trainer/policy/normal/std Std 0.644187 +trainer/policy/normal/std Max 4.97909 +trainer/policy/normal/std Min 0.336494 +trainer/policy/normal/log_std Mean 1.01788 +trainer/policy/normal/log_std Std 0.303128 +trainer/policy/normal/log_std Max 1.60525 +trainer/policy/normal/log_std Min -1.08918 +eval/num steps total 456385 +eval/num paths total 460 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00588105 +eval/Actions Std 0.983279 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43986 +time/logging (s) 0.00369414 +time/sampling batch (s) 0.279612 +time/saving (s) 0.0034746 +time/training (s) 6.28851 +time/epoch (s) 9.01515 +time/total (s) 4718.79 +Epoch -541 +---------------------------------- --------------- +2022-05-10 14:29:30.607679 PDT | [0] Epoch -540 finished +---------------------------------- --------------- +epoch -540 +replay_buffer/size 999033 +trainer/num train calls 461000 +trainer/Policy Loss -19.454 +trainer/Log Pis Mean 24.432 +trainer/Log Pis Std 13.7933 +trainer/Log Pis Max 69.3918 +trainer/Log Pis Min -7.22952 +trainer/policy/mean Mean -0.0437018 +trainer/policy/mean Std 0.90129 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.89011 +trainer/policy/normal/std Std 0.65436 +trainer/policy/normal/std Max 5.73965 +trainer/policy/normal/std Min 0.34573 +trainer/policy/normal/log_std Mean 1.02505 +trainer/policy/normal/log_std Std 0.301837 +trainer/policy/normal/log_std Max 1.7474 +trainer/policy/normal/log_std Min -1.0621 +eval/num steps total 457385 +eval/num paths total 461 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0535832 +eval/Actions Std 0.94952 +eval/Actions Max 0.999996 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.92703 +time/logging (s) 0.00381482 +time/sampling batch (s) 0.281504 +time/saving (s) 0.00367122 +time/training (s) 7.29603 +time/epoch (s) 10.512 +time/total (s) 4729.3 +Epoch -540 +---------------------------------- --------------- +2022-05-10 14:29:41.035816 PDT | [0] Epoch -539 finished +---------------------------------- --------------- +epoch -539 +replay_buffer/size 999033 +trainer/num train calls 462000 +trainer/Policy Loss -19.3633 +trainer/Log Pis Mean 23.9545 +trainer/Log Pis Std 12.9938 +trainer/Log Pis Max 63.764 +trainer/Log Pis Min -5.43264 +trainer/policy/mean Mean -0.0292304 +trainer/policy/mean Std 0.906806 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83792 +trainer/policy/normal/std Std 0.660743 +trainer/policy/normal/std Max 5.55233 +trainer/policy/normal/std Min 0.283496 +trainer/policy/normal/log_std Mean 1.00263 +trainer/policy/normal/log_std Std 0.323811 +trainer/policy/normal/log_std Max 1.71422 +trainer/policy/normal/log_std Min -1.26056 +eval/num steps total 458385 +eval/num paths total 462 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0843196 +eval/Actions Std 0.89189 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73499 +time/logging (s) 0.00396181 +time/sampling batch (s) 0.282181 +time/saving (s) 0.0039913 +time/training (s) 7.38168 +time/epoch (s) 10.4068 +time/total (s) 4739.71 +Epoch -539 +---------------------------------- --------------- +2022-05-10 14:29:51.133974 PDT | [0] Epoch -538 finished +---------------------------------- --------------- +epoch -538 +replay_buffer/size 999033 +trainer/num train calls 463000 +trainer/Policy Loss -19.2664 +trainer/Log Pis Mean 24.3129 +trainer/Log Pis Std 13.1422 +trainer/Log Pis Max 71.5477 +trainer/Log Pis Min -12.1584 +trainer/policy/mean Mean -0.0431526 +trainer/policy/mean Std 0.901811 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.73976 +trainer/policy/normal/std Std 0.641251 +trainer/policy/normal/std Max 7.06957 +trainer/policy/normal/std Min 0.32432 +trainer/policy/normal/log_std Mean 0.968082 +trainer/policy/normal/log_std Std 0.319005 +trainer/policy/normal/log_std Max 1.9558 +trainer/policy/normal/log_std Min -1.12602 +eval/num steps total 459385 +eval/num paths total 463 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.332935 +eval/Actions Std 0.832533 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83969 +time/logging (s) 0.00375089 +time/sampling batch (s) 0.28378 +time/saving (s) 0.00358143 +time/training (s) 6.9455 +time/epoch (s) 10.0763 +time/total (s) 4749.79 +Epoch -538 +---------------------------------- --------------- +2022-05-10 14:30:00.852100 PDT | [0] Epoch -537 finished +---------------------------------- --------------- +epoch -537 +replay_buffer/size 999033 +trainer/num train calls 464000 +trainer/Policy Loss -19.8228 +trainer/Log Pis Mean 23.1908 +trainer/Log Pis Std 13.472 +trainer/Log Pis Max 65.2432 +trainer/Log Pis Min -10.4371 +trainer/policy/mean Mean -0.0365738 +trainer/policy/mean Std 0.901546 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.86197 +trainer/policy/normal/std Std 0.664256 +trainer/policy/normal/std Max 6.21623 +trainer/policy/normal/std Min 0.308002 +trainer/policy/normal/log_std Mean 1.01337 +trainer/policy/normal/log_std Std 0.310588 +trainer/policy/normal/log_std Max 1.82716 +trainer/policy/normal/log_std Min -1.17765 +eval/num steps total 460385 +eval/num paths total 464 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118479 +eval/Actions Std 0.928815 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48526 +time/logging (s) 0.00378865 +time/sampling batch (s) 0.282655 +time/saving (s) 0.00348514 +time/training (s) 6.92161 +time/epoch (s) 9.69679 +time/total (s) 4759.49 +Epoch -537 +---------------------------------- --------------- +2022-05-10 14:30:11.002945 PDT | [0] Epoch -536 finished +---------------------------------- --------------- +epoch -536 +replay_buffer/size 999033 +trainer/num train calls 465000 +trainer/Policy Loss -18.8604 +trainer/Log Pis Mean 24.6243 +trainer/Log Pis Std 13.0695 +trainer/Log Pis Max 62.6154 +trainer/Log Pis Min -5.32915 +trainer/policy/mean Mean -0.0177528 +trainer/policy/mean Std 0.906643 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79316 +trainer/policy/normal/std Std 0.640355 +trainer/policy/normal/std Max 5.47138 +trainer/policy/normal/std Min 0.315391 +trainer/policy/normal/log_std Mean 0.988417 +trainer/policy/normal/log_std Std 0.316223 +trainer/policy/normal/log_std Max 1.69953 +trainer/policy/normal/log_std Min -1.15394 +eval/num steps total 461385 +eval/num paths total 465 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0563088 +eval/Actions Std 0.915255 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67441 +time/logging (s) 0.00387352 +time/sampling batch (s) 0.529921 +time/saving (s) 0.00369234 +time/training (s) 6.91753 +time/epoch (s) 10.1294 +time/total (s) 4769.62 +Epoch -536 +---------------------------------- --------------- +2022-05-10 14:30:21.049608 PDT | [0] Epoch -535 finished +---------------------------------- ---------------- +epoch -535 +replay_buffer/size 999033 +trainer/num train calls 466000 +trainer/Policy Loss -19.2492 +trainer/Log Pis Mean 24.7366 +trainer/Log Pis Std 13.6695 +trainer/Log Pis Max 85.6195 +trainer/Log Pis Min -5.57647 +trainer/policy/mean Mean -0.0688973 +trainer/policy/mean Std 0.903468 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8179 +trainer/policy/normal/std Std 0.630581 +trainer/policy/normal/std Max 5.54708 +trainer/policy/normal/std Min 0.274205 +trainer/policy/normal/log_std Mean 1.00165 +trainer/policy/normal/log_std Std 0.290926 +trainer/policy/normal/log_std Max 1.71327 +trainer/policy/normal/log_std Min -1.29388 +eval/num steps total 462385 +eval/num paths total 466 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000818272 +eval/Actions Std 0.904659 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68022 +time/logging (s) 0.00376075 +time/sampling batch (s) 0.278162 +time/saving (s) 0.00338097 +time/training (s) 6.80942 +time/epoch (s) 9.77494 +time/total (s) 4779.65 +Epoch -535 +---------------------------------- ---------------- +2022-05-10 14:30:30.803362 PDT | [0] Epoch -534 finished +---------------------------------- --------------- +epoch -534 +replay_buffer/size 999033 +trainer/num train calls 467000 +trainer/Policy Loss -20.1498 +trainer/Log Pis Mean 25.3305 +trainer/Log Pis Std 13.6027 +trainer/Log Pis Max 68.3546 +trainer/Log Pis Min -4.46395 +trainer/policy/mean Mean -0.0252421 +trainer/policy/mean Std 0.902484 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.83882 +trainer/policy/normal/std Std 0.63619 +trainer/policy/normal/std Max 5.95314 +trainer/policy/normal/std Min 0.339251 +trainer/policy/normal/log_std Mean 1.00762 +trainer/policy/normal/log_std Std 0.300835 +trainer/policy/normal/log_std Max 1.78392 +trainer/policy/normal/log_std Min -1.08102 +eval/num steps total 463385 +eval/num paths total 467 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111127 +eval/Actions Std 0.895545 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64561 +time/logging (s) 0.00417295 +time/sampling batch (s) 0.278361 +time/saving (s) 0.00407275 +time/training (s) 6.80083 +time/epoch (s) 9.73305 +time/total (s) 4789.39 +Epoch -534 +---------------------------------- --------------- +2022-05-10 14:30:41.011208 PDT | [0] Epoch -533 finished +---------------------------------- --------------- +epoch -533 +replay_buffer/size 999033 +trainer/num train calls 468000 +trainer/Policy Loss -19.9367 +trainer/Log Pis Mean 24.4327 +trainer/Log Pis Std 13.2005 +trainer/Log Pis Max 71.716 +trainer/Log Pis Min -4.27255 +trainer/policy/mean Mean -0.0402592 +trainer/policy/mean Std 0.903864 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82016 +trainer/policy/normal/std Std 0.659088 +trainer/policy/normal/std Max 6.02607 +trainer/policy/normal/std Min 0.343186 +trainer/policy/normal/log_std Mean 0.997453 +trainer/policy/normal/log_std Std 0.316349 +trainer/policy/normal/log_std Max 1.7961 +trainer/policy/normal/log_std Min -1.06948 +eval/num steps total 464385 +eval/num paths total 468 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.350622 +eval/Actions Std 0.841406 +eval/Actions Max 0.999991 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48919 +time/logging (s) 0.00375558 +time/sampling batch (s) 0.528379 +time/saving (s) 0.00369247 +time/training (s) 7.1609 +time/epoch (s) 10.1859 +time/total (s) 4799.58 +Epoch -533 +---------------------------------- --------------- +2022-05-10 14:30:51.146553 PDT | [0] Epoch -532 finished +---------------------------------- --------------- +epoch -532 +replay_buffer/size 999033 +trainer/num train calls 469000 +trainer/Policy Loss -18.3145 +trainer/Log Pis Mean 24.2704 +trainer/Log Pis Std 13.6933 +trainer/Log Pis Max 76.6002 +trainer/Log Pis Min -6.7622 +trainer/policy/mean Mean -0.0420146 +trainer/policy/mean Std 0.909867 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79073 +trainer/policy/normal/std Std 0.63225 +trainer/policy/normal/std Max 6.96305 +trainer/policy/normal/std Min 0.345558 +trainer/policy/normal/log_std Mean 0.990328 +trainer/policy/normal/log_std Std 0.301118 +trainer/policy/normal/log_std Max 1.94062 +trainer/policy/normal/log_std Min -1.06259 +eval/num steps total 465385 +eval/num paths total 469 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122848 +eval/Actions Std 0.837055 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6545 +time/logging (s) 0.00366889 +time/sampling batch (s) 0.278851 +time/saving (s) 0.00345574 +time/training (s) 7.17344 +time/epoch (s) 10.1139 +time/total (s) 4809.69 +Epoch -532 +---------------------------------- --------------- +2022-05-10 14:31:01.014940 PDT | [0] Epoch -531 finished +---------------------------------- --------------- +epoch -531 +replay_buffer/size 999033 +trainer/num train calls 470000 +trainer/Policy Loss -20.0603 +trainer/Log Pis Mean 24.3392 +trainer/Log Pis Std 13.6294 +trainer/Log Pis Max 68.9894 +trainer/Log Pis Min -7.36115 +trainer/policy/mean Mean -0.0397932 +trainer/policy/mean Std 0.904671 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85101 +trainer/policy/normal/std Std 0.626427 +trainer/policy/normal/std Max 5.8021 +trainer/policy/normal/std Min 0.300756 +trainer/policy/normal/log_std Mean 1.01349 +trainer/policy/normal/log_std Std 0.294539 +trainer/policy/normal/log_std Max 1.75822 +trainer/policy/normal/log_std Min -1.20146 +eval/num steps total 466385 +eval/num paths total 470 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0156694 +eval/Actions Std 0.900506 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33409 +time/logging (s) 0.00381313 +time/sampling batch (s) 0.277979 +time/saving (s) 0.00342125 +time/training (s) 7.22821 +time/epoch (s) 9.8475 +time/total (s) 4819.54 +Epoch -531 +---------------------------------- --------------- +2022-05-10 14:31:11.023062 PDT | [0] Epoch -530 finished +---------------------------------- --------------- +epoch -530 +replay_buffer/size 999033 +trainer/num train calls 471000 +trainer/Policy Loss -20.2779 +trainer/Log Pis Mean 25.8621 +trainer/Log Pis Std 13.5433 +trainer/Log Pis Max 76.4731 +trainer/Log Pis Min -7.32537 +trainer/policy/mean Mean -0.0352438 +trainer/policy/mean Std 0.908656 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84854 +trainer/policy/normal/std Std 0.645619 +trainer/policy/normal/std Max 5.34446 +trainer/policy/normal/std Min 0.269504 +trainer/policy/normal/log_std Mean 1.0093 +trainer/policy/normal/log_std Std 0.309814 +trainer/policy/normal/log_std Max 1.67606 +trainer/policy/normal/log_std Min -1.31117 +eval/num steps total 467385 +eval/num paths total 471 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.103788 +eval/Actions Std 0.912915 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6221 +time/logging (s) 0.00375753 +time/sampling batch (s) 0.278374 +time/saving (s) 0.00354911 +time/training (s) 7.07916 +time/epoch (s) 9.98694 +time/total (s) 4829.53 +Epoch -530 +---------------------------------- --------------- +2022-05-10 14:31:21.759133 PDT | [0] Epoch -529 finished +---------------------------------- --------------- +epoch -529 +replay_buffer/size 999033 +trainer/num train calls 472000 +trainer/Policy Loss -18.0766 +trainer/Log Pis Mean 24.8352 +trainer/Log Pis Std 13.1058 +trainer/Log Pis Max 68.3615 +trainer/Log Pis Min -4.46734 +trainer/policy/mean Mean -0.0463871 +trainer/policy/mean Std 0.898674 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8055 +trainer/policy/normal/std Std 0.665787 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.339277 +trainer/policy/normal/log_std Mean 0.989497 +trainer/policy/normal/log_std Std 0.330566 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.08094 +eval/num steps total 468385 +eval/num paths total 472 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.02237 +eval/Actions Std 0.919926 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65171 +time/logging (s) 0.00367059 +time/sampling batch (s) 0.530889 +time/saving (s) 0.00345005 +time/training (s) 7.52503 +time/epoch (s) 10.7147 +time/total (s) 4840.25 +Epoch -529 +---------------------------------- --------------- +2022-05-10 14:31:31.887412 PDT | [0] Epoch -528 finished +---------------------------------- --------------- +epoch -528 +replay_buffer/size 999033 +trainer/num train calls 473000 +trainer/Policy Loss -19.659 +trainer/Log Pis Mean 24.1562 +trainer/Log Pis Std 13.9964 +trainer/Log Pis Max 70.5145 +trainer/Log Pis Min -7.31735 +trainer/policy/mean Mean -0.0283808 +trainer/policy/mean Std 0.912781 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81355 +trainer/policy/normal/std Std 0.655873 +trainer/policy/normal/std Max 6.34854 +trainer/policy/normal/std Min 0.320585 +trainer/policy/normal/log_std Mean 0.994681 +trainer/policy/normal/log_std Std 0.319736 +trainer/policy/normal/log_std Max 1.84823 +trainer/policy/normal/log_std Min -1.13761 +eval/num steps total 469385 +eval/num paths total 473 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.428951 +eval/Actions Std 0.854514 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54857 +time/logging (s) 0.00420405 +time/sampling batch (s) 0.28443 +time/saving (s) 0.00408721 +time/training (s) 7.26606 +time/epoch (s) 10.1074 +time/total (s) 4850.36 +Epoch -528 +---------------------------------- --------------- +2022-05-10 14:31:42.761140 PDT | [0] Epoch -527 finished +---------------------------------- -------------- +epoch -527 +replay_buffer/size 999033 +trainer/num train calls 474000 +trainer/Policy Loss -20.366 +trainer/Log Pis Mean 24.6358 +trainer/Log Pis Std 13.684 +trainer/Log Pis Max 72.8491 +trainer/Log Pis Min -11.0015 +trainer/policy/mean Mean -0.0363076 +trainer/policy/mean Std 0.908694 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82152 +trainer/policy/normal/std Std 0.642404 +trainer/policy/normal/std Max 5.41594 +trainer/policy/normal/std Min 0.319336 +trainer/policy/normal/log_std Mean 0.998947 +trainer/policy/normal/log_std Std 0.314203 +trainer/policy/normal/log_std Max 1.68935 +trainer/policy/normal/log_std Min -1.14151 +eval/num steps total 470385 +eval/num paths total 474 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0084948 +eval/Actions Std 0.911056 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72109 +time/logging (s) 0.0037087 +time/sampling batch (s) 0.28202 +time/saving (s) 0.0035711 +time/training (s) 7.84126 +time/epoch (s) 10.8516 +time/total (s) 4861.22 +Epoch -527 +---------------------------------- -------------- +2022-05-10 14:31:53.350838 PDT | [0] Epoch -526 finished +---------------------------------- --------------- +epoch -526 +replay_buffer/size 999033 +trainer/num train calls 475000 +trainer/Policy Loss -20.2164 +trainer/Log Pis Mean 25.1986 +trainer/Log Pis Std 13.2133 +trainer/Log Pis Max 61.1975 +trainer/Log Pis Min -7.24993 +trainer/policy/mean Mean -0.016354 +trainer/policy/mean Std 0.91092 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82608 +trainer/policy/normal/std Std 0.642278 +trainer/policy/normal/std Max 5.05116 +trainer/policy/normal/std Min 0.312509 +trainer/policy/normal/log_std Mean 1.00137 +trainer/policy/normal/log_std Std 0.310252 +trainer/policy/normal/log_std Max 1.61962 +trainer/policy/normal/log_std Min -1.16312 +eval/num steps total 471385 +eval/num paths total 475 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0807562 +eval/Actions Std 0.911444 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81969 +time/logging (s) 0.00370589 +time/sampling batch (s) 0.282772 +time/saving (s) 0.00343975 +time/training (s) 7.45854 +time/epoch (s) 10.5681 +time/total (s) 4871.79 +Epoch -526 +---------------------------------- --------------- +2022-05-10 14:32:02.644464 PDT | [0] Epoch -525 finished +---------------------------------- --------------- +epoch -525 +replay_buffer/size 999033 +trainer/num train calls 476000 +trainer/Policy Loss -19.6598 +trainer/Log Pis Mean 24.2446 +trainer/Log Pis Std 13.1575 +trainer/Log Pis Max 69.7298 +trainer/Log Pis Min -11.0753 +trainer/policy/mean Mean -0.0298598 +trainer/policy/mean Std 0.904873 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86097 +trainer/policy/normal/std Std 0.650303 +trainer/policy/normal/std Max 5.04237 +trainer/policy/normal/std Min 0.276779 +trainer/policy/normal/log_std Mean 1.01293 +trainer/policy/normal/log_std Std 0.314048 +trainer/policy/normal/log_std Max 1.61788 +trainer/policy/normal/log_std Min -1.28454 +eval/num steps total 472385 +eval/num paths total 476 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0284995 +eval/Actions Std 0.909854 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67324 +time/logging (s) 0.0036607 +time/sampling batch (s) 0.280182 +time/saving (s) 0.00337858 +time/training (s) 6.31187 +time/epoch (s) 9.27233 +time/total (s) 4881.06 +Epoch -525 +---------------------------------- --------------- +2022-05-10 14:32:12.538756 PDT | [0] Epoch -524 finished +---------------------------------- --------------- +epoch -524 +replay_buffer/size 999033 +trainer/num train calls 477000 +trainer/Policy Loss -19.4694 +trainer/Log Pis Mean 24.3274 +trainer/Log Pis Std 13.2782 +trainer/Log Pis Max 64.6438 +trainer/Log Pis Min -8.67566 +trainer/policy/mean Mean -0.0270866 +trainer/policy/mean Std 0.910081 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80643 +trainer/policy/normal/std Std 0.612798 +trainer/policy/normal/std Max 5.94385 +trainer/policy/normal/std Min 0.332977 +trainer/policy/normal/log_std Mean 0.998602 +trainer/policy/normal/log_std Std 0.287803 +trainer/policy/normal/log_std Max 1.78236 +trainer/policy/normal/log_std Min -1.09968 +eval/num steps total 473385 +eval/num paths total 477 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0584013 +eval/Actions Std 0.908678 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64067 +time/logging (s) 0.00368745 +time/sampling batch (s) 0.283071 +time/saving (s) 0.00346443 +time/training (s) 6.94197 +time/epoch (s) 9.87286 +time/total (s) 4890.94 +Epoch -524 +---------------------------------- --------------- +2022-05-10 14:32:23.152462 PDT | [0] Epoch -523 finished +---------------------------------- --------------- +epoch -523 +replay_buffer/size 999033 +trainer/num train calls 478000 +trainer/Policy Loss -18.7159 +trainer/Log Pis Mean 25.2098 +trainer/Log Pis Std 13.3486 +trainer/Log Pis Max 65.8117 +trainer/Log Pis Min -8.3139 +trainer/policy/mean Mean -0.0158701 +trainer/policy/mean Std 0.906063 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8285 +trainer/policy/normal/std Std 0.639126 +trainer/policy/normal/std Max 5.37919 +trainer/policy/normal/std Min 0.306677 +trainer/policy/normal/log_std Mean 1.00413 +trainer/policy/normal/log_std Std 0.297976 +trainer/policy/normal/log_std Max 1.68254 +trainer/policy/normal/log_std Min -1.18196 +eval/num steps total 474385 +eval/num paths total 478 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.384942 +eval/Actions Std 0.838489 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44659 +time/logging (s) 0.00366883 +time/sampling batch (s) 0.283647 +time/saving (s) 0.00341977 +time/training (s) 7.85471 +time/epoch (s) 10.592 +time/total (s) 4901.53 +Epoch -523 +---------------------------------- --------------- +2022-05-10 14:32:34.188896 PDT | [0] Epoch -522 finished +---------------------------------- -------------- +epoch -522 +replay_buffer/size 999033 +trainer/num train calls 479000 +trainer/Policy Loss -20.9745 +trainer/Log Pis Mean 25.1838 +trainer/Log Pis Std 13.9076 +trainer/Log Pis Max 73.1046 +trainer/Log Pis Min -7.72135 +trainer/policy/mean Mean -0.0659731 +trainer/policy/mean Std 0.908306 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82338 +trainer/policy/normal/std Std 0.633364 +trainer/policy/normal/std Max 5.89256 +trainer/policy/normal/std Min 0.347726 +trainer/policy/normal/log_std Mean 1.00236 +trainer/policy/normal/log_std Std 0.298528 +trainer/policy/normal/log_std Max 1.77369 +trainer/policy/normal/log_std Min -1.05634 +eval/num steps total 475385 +eval/num paths total 479 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.289683 +eval/Actions Std 0.883857 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61288 +time/logging (s) 0.0041094 +time/sampling batch (s) 0.289049 +time/saving (s) 0.0042315 +time/training (s) 8.10479 +time/epoch (s) 11.0151 +time/total (s) 4912.55 +Epoch -522 +---------------------------------- -------------- +2022-05-10 14:32:43.313297 PDT | [0] Epoch -521 finished +---------------------------------- --------------- +epoch -521 +replay_buffer/size 999033 +trainer/num train calls 480000 +trainer/Policy Loss -18.8462 +trainer/Log Pis Mean 23.6776 +trainer/Log Pis Std 12.5143 +trainer/Log Pis Max 60.2135 +trainer/Log Pis Min -8.12179 +trainer/policy/mean Mean -0.0384521 +trainer/policy/mean Std 0.905417 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.75679 +trainer/policy/normal/std Std 0.621262 +trainer/policy/normal/std Max 5.51468 +trainer/policy/normal/std Min 0.323435 +trainer/policy/normal/log_std Mean 0.977636 +trainer/policy/normal/log_std Std 0.303479 +trainer/policy/normal/log_std Max 1.70741 +trainer/policy/normal/log_std Min -1.12876 +eval/num steps total 476385 +eval/num paths total 480 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.400436 +eval/Actions Std 0.870501 +eval/Actions Max 0.999973 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82763 +time/logging (s) 0.00387662 +time/sampling batch (s) 0.290209 +time/saving (s) 0.00349361 +time/training (s) 5.97638 +time/epoch (s) 9.10159 +time/total (s) 4921.65 +Epoch -521 +---------------------------------- --------------- +2022-05-10 14:32:54.053315 PDT | [0] Epoch -520 finished +---------------------------------- --------------- +epoch -520 +replay_buffer/size 999033 +trainer/num train calls 481000 +trainer/Policy Loss -19.7431 +trainer/Log Pis Mean 24.8991 +trainer/Log Pis Std 13.1426 +trainer/Log Pis Max 79.9824 +trainer/Log Pis Min -7.73674 +trainer/policy/mean Mean -0.0263599 +trainer/policy/mean Std 0.909159 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84759 +trainer/policy/normal/std Std 0.638516 +trainer/policy/normal/std Max 5.8866 +trainer/policy/normal/std Min 0.290617 +trainer/policy/normal/log_std Mean 1.00926 +trainer/policy/normal/log_std Std 0.30965 +trainer/policy/normal/log_std Max 1.77268 +trainer/policy/normal/log_std Min -1.23575 +eval/num steps total 477385 +eval/num paths total 481 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138672 +eval/Actions Std 0.918094 +eval/Actions Max 0.999993 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63405 +time/logging (s) 0.00528673 +time/sampling batch (s) 0.544484 +time/saving (s) 0.00356101 +time/training (s) 7.53179 +time/epoch (s) 10.7192 +time/total (s) 4932.38 +Epoch -520 +---------------------------------- --------------- +2022-05-10 14:33:03.767528 PDT | [0] Epoch -519 finished +---------------------------------- --------------- +epoch -519 +replay_buffer/size 999033 +trainer/num train calls 482000 +trainer/Policy Loss -19.5524 +trainer/Log Pis Mean 24.6143 +trainer/Log Pis Std 12.5342 +trainer/Log Pis Max 62.9058 +trainer/Log Pis Min -6.21862 +trainer/policy/mean Mean -0.0476091 +trainer/policy/mean Std 0.906327 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83025 +trainer/policy/normal/std Std 0.641829 +trainer/policy/normal/std Max 5.83817 +trainer/policy/normal/std Min 0.315192 +trainer/policy/normal/log_std Mean 1.00418 +trainer/policy/normal/log_std Std 0.301056 +trainer/policy/normal/log_std Max 1.76442 +trainer/policy/normal/log_std Min -1.15457 +eval/num steps total 478385 +eval/num paths total 482 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00537666 +eval/Actions Std 0.981838 +eval/Actions Max 0.99995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71667 +time/logging (s) 0.0038232 +time/sampling batch (s) 0.292684 +time/saving (s) 0.00351763 +time/training (s) 6.67368 +time/epoch (s) 9.69037 +time/total (s) 4942.07 +Epoch -519 +---------------------------------- --------------- +2022-05-10 14:33:13.425586 PDT | [0] Epoch -518 finished +---------------------------------- --------------- +epoch -518 +replay_buffer/size 999033 +trainer/num train calls 483000 +trainer/Policy Loss -19.8179 +trainer/Log Pis Mean 22.9786 +trainer/Log Pis Std 12.7741 +trainer/Log Pis Max 71.5958 +trainer/Log Pis Min -8.02973 +trainer/policy/mean Mean -0.0257923 +trainer/policy/mean Std 0.901175 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77613 +trainer/policy/normal/std Std 0.655239 +trainer/policy/normal/std Max 5.51647 +trainer/policy/normal/std Min 0.339037 +trainer/policy/normal/log_std Mean 0.98024 +trainer/policy/normal/log_std Std 0.323316 +trainer/policy/normal/log_std Max 1.70774 +trainer/policy/normal/log_std Min -1.08165 +eval/num steps total 479385 +eval/num paths total 483 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.342737 +eval/Actions Std 0.832849 +eval/Actions Max 0.999994 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7389 +time/logging (s) 0.00381289 +time/sampling batch (s) 0.541071 +time/saving (s) 0.00361339 +time/training (s) 6.34844 +time/epoch (s) 9.63584 +time/total (s) 4951.71 +Epoch -518 +---------------------------------- --------------- +2022-05-10 14:33:23.257993 PDT | [0] Epoch -517 finished +---------------------------------- --------------- +epoch -517 +replay_buffer/size 999033 +trainer/num train calls 484000 +trainer/Policy Loss -20.8122 +trainer/Log Pis Mean 24.5829 +trainer/Log Pis Std 13.0589 +trainer/Log Pis Max 67.9904 +trainer/Log Pis Min -10.6285 +trainer/policy/mean Mean -0.0259954 +trainer/policy/mean Std 0.905644 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83341 +trainer/policy/normal/std Std 0.637873 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.35889 +trainer/policy/normal/log_std Mean 1.00534 +trainer/policy/normal/log_std Std 0.303596 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.02474 +eval/num steps total 480385 +eval/num paths total 484 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.264625 +eval/Actions Std 0.892422 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69632 +time/logging (s) 0.00417923 +time/sampling batch (s) 0.282908 +time/saving (s) 0.00416137 +time/training (s) 6.82318 +time/epoch (s) 9.81075 +time/total (s) 4961.52 +Epoch -517 +---------------------------------- --------------- +2022-05-10 14:33:34.127425 PDT | [0] Epoch -516 finished +---------------------------------- --------------- +epoch -516 +replay_buffer/size 999033 +trainer/num train calls 485000 +trainer/Policy Loss -20.0241 +trainer/Log Pis Mean 25.2367 +trainer/Log Pis Std 13.088 +trainer/Log Pis Max 76.845 +trainer/Log Pis Min -6.40812 +trainer/policy/mean Mean -0.046717 +trainer/policy/mean Std 0.907567 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82766 +trainer/policy/normal/std Std 0.661936 +trainer/policy/normal/std Max 6.33533 +trainer/policy/normal/std Min 0.328885 +trainer/policy/normal/log_std Mean 0.999138 +trainer/policy/normal/log_std Std 0.322212 +trainer/policy/normal/log_std Max 1.84614 +trainer/policy/normal/log_std Min -1.11205 +eval/num steps total 481385 +eval/num paths total 485 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.19067 +eval/Actions Std 0.941517 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75413 +time/logging (s) 0.00384428 +time/sampling batch (s) 0.282805 +time/saving (s) 0.00360621 +time/training (s) 7.80267 +time/epoch (s) 10.8471 +time/total (s) 4972.37 +Epoch -516 +---------------------------------- --------------- +2022-05-10 14:33:44.056825 PDT | [0] Epoch -515 finished +---------------------------------- --------------- +epoch -515 +replay_buffer/size 999033 +trainer/num train calls 486000 +trainer/Policy Loss -19.018 +trainer/Log Pis Mean 24.3844 +trainer/Log Pis Std 12.9382 +trainer/Log Pis Max 63.9087 +trainer/Log Pis Min -5.31405 +trainer/policy/mean Mean -0.0457597 +trainer/policy/mean Std 0.900853 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78168 +trainer/policy/normal/std Std 0.646413 +trainer/policy/normal/std Max 6.47138 +trainer/policy/normal/std Min 0.31186 +trainer/policy/normal/log_std Mean 0.984273 +trainer/policy/normal/log_std Std 0.314361 +trainer/policy/normal/log_std Max 1.86739 +trainer/policy/normal/log_std Min -1.1652 +eval/num steps total 482385 +eval/num paths total 486 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.326471 +eval/Actions Std 0.820789 +eval/Actions Max 0.99999 +eval/Actions Min -0.999973 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67946 +time/logging (s) 0.00372671 +time/sampling batch (s) 0.285089 +time/saving (s) 0.00343439 +time/training (s) 6.93573 +time/epoch (s) 9.90744 +time/total (s) 4982.28 +Epoch -515 +---------------------------------- --------------- +2022-05-10 14:33:53.898368 PDT | [0] Epoch -514 finished +---------------------------------- --------------- +epoch -514 +replay_buffer/size 999033 +trainer/num train calls 487000 +trainer/Policy Loss -19.6196 +trainer/Log Pis Mean 24.7547 +trainer/Log Pis Std 12.9255 +trainer/Log Pis Max 62.1504 +trainer/Log Pis Min -7.0309 +trainer/policy/mean Mean -0.0497538 +trainer/policy/mean Std 0.906147 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80103 +trainer/policy/normal/std Std 0.635555 +trainer/policy/normal/std Max 5.75461 +trainer/policy/normal/std Min 0.354975 +trainer/policy/normal/log_std Mean 0.99256 +trainer/policy/normal/log_std Std 0.309654 +trainer/policy/normal/log_std Max 1.75 +trainer/policy/normal/log_std Min -1.03571 +eval/num steps total 483385 +eval/num paths total 487 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.285861 +eval/Actions Std 0.851676 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39036 +time/logging (s) 0.00420271 +time/sampling batch (s) 0.546772 +time/saving (s) 0.00354898 +time/training (s) 6.8749 +time/epoch (s) 9.81978 +time/total (s) 4992.11 +Epoch -514 +---------------------------------- --------------- +2022-05-10 14:34:05.215103 PDT | [0] Epoch -513 finished +---------------------------------- --------------- +epoch -513 +replay_buffer/size 999033 +trainer/num train calls 488000 +trainer/Policy Loss -19.3328 +trainer/Log Pis Mean 24.208 +trainer/Log Pis Std 12.996 +trainer/Log Pis Max 61.4809 +trainer/Log Pis Min -3.90594 +trainer/policy/mean Mean -0.0549301 +trainer/policy/mean Std 0.905457 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.85494 +trainer/policy/normal/std Std 0.624253 +trainer/policy/normal/std Max 4.98985 +trainer/policy/normal/std Min 0.262015 +trainer/policy/normal/log_std Mean 1.01582 +trainer/policy/normal/log_std Std 0.287292 +trainer/policy/normal/log_std Max 1.60741 +trainer/policy/normal/log_std Min -1.33935 +eval/num steps total 484385 +eval/num paths total 488 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0621109 +eval/Actions Std 0.893085 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87615 +time/logging (s) 0.00424363 +time/sampling batch (s) 0.2805 +time/saving (s) 0.00421674 +time/training (s) 8.12949 +time/epoch (s) 11.2946 +time/total (s) 5003.4 +Epoch -513 +---------------------------------- --------------- +2022-05-10 14:34:16.233549 PDT | [0] Epoch -512 finished +---------------------------------- --------------- +epoch -512 +replay_buffer/size 999033 +trainer/num train calls 489000 +trainer/Policy Loss -19.899 +trainer/Log Pis Mean 24.5097 +trainer/Log Pis Std 13.0084 +trainer/Log Pis Max 69.4327 +trainer/Log Pis Min -5.83454 +trainer/policy/mean Mean -0.0438427 +trainer/policy/mean Std 0.9069 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79577 +trainer/policy/normal/std Std 0.603951 +trainer/policy/normal/std Max 6.4492 +trainer/policy/normal/std Min 0.29902 +trainer/policy/normal/log_std Mean 0.994284 +trainer/policy/normal/log_std Std 0.294285 +trainer/policy/normal/log_std Max 1.86396 +trainer/policy/normal/log_std Min -1.20725 +eval/num steps total 485385 +eval/num paths total 489 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.183149 +eval/Actions Std 0.873546 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58296 +time/logging (s) 0.00375261 +time/sampling batch (s) 0.530353 +time/saving (s) 0.00358099 +time/training (s) 7.87522 +time/epoch (s) 10.9959 +time/total (s) 5014.4 +Epoch -512 +---------------------------------- --------------- +2022-05-10 14:34:26.103695 PDT | [0] Epoch -511 finished +---------------------------------- --------------- +epoch -511 +replay_buffer/size 999033 +trainer/num train calls 490000 +trainer/Policy Loss -20.3198 +trainer/Log Pis Mean 23.8588 +trainer/Log Pis Std 12.9435 +trainer/Log Pis Max 66.5504 +trainer/Log Pis Min -11.0877 +trainer/policy/mean Mean -0.0396928 +trainer/policy/mean Std 0.901512 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79202 +trainer/policy/normal/std Std 0.659077 +trainer/policy/normal/std Max 4.97675 +trainer/policy/normal/std Min 0.319311 +trainer/policy/normal/log_std Mean 0.986136 +trainer/policy/normal/log_std Std 0.321814 +trainer/policy/normal/log_std Max 1.60478 +trainer/policy/normal/log_std Min -1.14159 +eval/num steps total 486385 +eval/num paths total 490 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0835764 +eval/Actions Std 0.915579 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60269 +time/logging (s) 0.00367743 +time/sampling batch (s) 0.280961 +time/saving (s) 0.00341395 +time/training (s) 6.95774 +time/epoch (s) 9.84848 +time/total (s) 5024.25 +Epoch -511 +---------------------------------- --------------- +2022-05-10 14:34:36.714339 PDT | [0] Epoch -510 finished +---------------------------------- --------------- +epoch -510 +replay_buffer/size 999033 +trainer/num train calls 491000 +trainer/Policy Loss -19.8413 +trainer/Log Pis Mean 25.7946 +trainer/Log Pis Std 12.9173 +trainer/Log Pis Max 74.0192 +trainer/Log Pis Min -7.77145 +trainer/policy/mean Mean -0.0218087 +trainer/policy/mean Std 0.907277 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.82503 +trainer/policy/normal/std Std 0.617037 +trainer/policy/normal/std Max 5.27101 +trainer/policy/normal/std Min 0.366938 +trainer/policy/normal/log_std Mean 1.00531 +trainer/policy/normal/log_std Std 0.286614 +trainer/policy/normal/log_std Max 1.66222 +trainer/policy/normal/log_std Min -1.00256 +eval/num steps total 487385 +eval/num paths total 491 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0302206 +eval/Actions Std 0.954716 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6717 +time/logging (s) 0.00436287 +time/sampling batch (s) 0.532187 +time/saving (s) 0.00429692 +time/training (s) 7.37717 +time/epoch (s) 10.5897 +time/total (s) 5034.85 +Epoch -510 +---------------------------------- --------------- +2022-05-10 14:34:46.846114 PDT | [0] Epoch -509 finished +---------------------------------- --------------- +epoch -509 +replay_buffer/size 999033 +trainer/num train calls 492000 +trainer/Policy Loss -18.4702 +trainer/Log Pis Mean 23.5617 +trainer/Log Pis Std 12.9381 +trainer/Log Pis Max 84.2481 +trainer/Log Pis Min -7.13929 +trainer/policy/mean Mean -0.0363 +trainer/policy/mean Std 0.907343 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75798 +trainer/policy/normal/std Std 0.628339 +trainer/policy/normal/std Max 5.1051 +trainer/policy/normal/std Min 0.305675 +trainer/policy/normal/log_std Mean 0.975956 +trainer/policy/normal/log_std Std 0.315682 +trainer/policy/normal/log_std Max 1.63024 +trainer/policy/normal/log_std Min -1.18523 +eval/num steps total 488385 +eval/num paths total 492 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107062 +eval/Actions Std 0.91145 +eval/Actions Max 0.999992 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69017 +time/logging (s) 0.00381314 +time/sampling batch (s) 0.281292 +time/saving (s) 0.00397537 +time/training (s) 7.12996 +time/epoch (s) 10.1092 +time/total (s) 5044.96 +Epoch -509 +---------------------------------- --------------- +2022-05-10 14:34:57.169151 PDT | [0] Epoch -508 finished +---------------------------------- --------------- +epoch -508 +replay_buffer/size 999033 +trainer/num train calls 493000 +trainer/Policy Loss -19.7166 +trainer/Log Pis Mean 24.2224 +trainer/Log Pis Std 13.0815 +trainer/Log Pis Max 68.9154 +trainer/Log Pis Min -5.87887 +trainer/policy/mean Mean -0.0219553 +trainer/policy/mean Std 0.906236 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.75819 +trainer/policy/normal/std Std 0.620249 +trainer/policy/normal/std Max 5.72022 +trainer/policy/normal/std Min 0.32162 +trainer/policy/normal/log_std Mean 0.978989 +trainer/policy/normal/log_std Std 0.298497 +trainer/policy/normal/log_std Max 1.74401 +trainer/policy/normal/log_std Min -1.13438 +eval/num steps total 489385 +eval/num paths total 493 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0270618 +eval/Actions Std 0.903498 +eval/Actions Max 0.999996 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70443 +time/logging (s) 0.0037657 +time/sampling batch (s) 0.288356 +time/saving (s) 0.00344713 +time/training (s) 7.30101 +time/epoch (s) 10.301 +time/total (s) 5055.26 +Epoch -508 +---------------------------------- --------------- +2022-05-10 14:35:07.465715 PDT | [0] Epoch -507 finished +---------------------------------- --------------- +epoch -507 +replay_buffer/size 999033 +trainer/num train calls 494000 +trainer/Policy Loss -19.5763 +trainer/Log Pis Mean 23.8706 +trainer/Log Pis Std 13.0619 +trainer/Log Pis Max 67.7889 +trainer/Log Pis Min -6.97622 +trainer/policy/mean Mean -0.0303085 +trainer/policy/mean Std 0.907695 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82608 +trainer/policy/normal/std Std 0.63877 +trainer/policy/normal/std Max 5.5363 +trainer/policy/normal/std Min 0.328742 +trainer/policy/normal/log_std Mean 1.00249 +trainer/policy/normal/log_std Std 0.303171 +trainer/policy/normal/log_std Max 1.71133 +trainer/policy/normal/log_std Min -1.11248 +eval/num steps total 490385 +eval/num paths total 494 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.019241 +eval/Actions Std 0.889305 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68484 +time/logging (s) 0.00377349 +time/sampling batch (s) 0.284654 +time/saving (s) 0.00342265 +time/training (s) 7.29809 +time/epoch (s) 10.2748 +time/total (s) 5065.54 +Epoch -507 +---------------------------------- --------------- +2022-05-10 14:35:18.391753 PDT | [0] Epoch -506 finished +---------------------------------- --------------- +epoch -506 +replay_buffer/size 999033 +trainer/num train calls 495000 +trainer/Policy Loss -19.2415 +trainer/Log Pis Mean 25.7552 +trainer/Log Pis Std 13.9982 +trainer/Log Pis Max 73.6141 +trainer/Log Pis Min -7.98634 +trainer/policy/mean Mean -0.0497073 +trainer/policy/mean Std 0.907908 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83589 +trainer/policy/normal/std Std 0.633808 +trainer/policy/normal/std Max 5.84091 +trainer/policy/normal/std Min 0.314558 +trainer/policy/normal/log_std Mean 1.00782 +trainer/policy/normal/log_std Std 0.292866 +trainer/policy/normal/log_std Max 1.76489 +trainer/policy/normal/log_std Min -1.15659 +eval/num steps total 491385 +eval/num paths total 495 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.178675 +eval/Actions Std 0.835672 +eval/Actions Max 0.999983 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60014 +time/logging (s) 0.00424632 +time/sampling batch (s) 0.535295 +time/saving (s) 0.00417358 +time/training (s) 7.76089 +time/epoch (s) 10.9048 +time/total (s) 5076.45 +Epoch -506 +---------------------------------- --------------- +2022-05-10 14:35:28.044579 PDT | [0] Epoch -505 finished +---------------------------------- --------------- +epoch -505 +replay_buffer/size 999033 +trainer/num train calls 496000 +trainer/Policy Loss -18.2171 +trainer/Log Pis Mean 24.7737 +trainer/Log Pis Std 13.2904 +trainer/Log Pis Max 71.0099 +trainer/Log Pis Min -7.26522 +trainer/policy/mean Mean -0.0274052 +trainer/policy/mean Std 0.905501 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999971 +trainer/policy/normal/std Mean 2.87434 +trainer/policy/normal/std Std 0.647556 +trainer/policy/normal/std Max 6.60921 +trainer/policy/normal/std Min 0.392494 +trainer/policy/normal/log_std Mean 1.02195 +trainer/policy/normal/log_std Std 0.287243 +trainer/policy/normal/log_std Max 1.88846 +trainer/policy/normal/log_std Min -0.935235 +eval/num steps total 492385 +eval/num paths total 496 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0138614 +eval/Actions Std 0.905047 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52621 +time/logging (s) 0.00379782 +time/sampling batch (s) 0.528584 +time/saving (s) 0.00354891 +time/training (s) 6.5685 +time/epoch (s) 9.63064 +time/total (s) 5086.08 +Epoch -505 +---------------------------------- --------------- +2022-05-10 14:35:37.016676 PDT | [0] Epoch -504 finished +---------------------------------- --------------- +epoch -504 +replay_buffer/size 999033 +trainer/num train calls 497000 +trainer/Policy Loss -20.4962 +trainer/Log Pis Mean 23.6852 +trainer/Log Pis Std 13.3806 +trainer/Log Pis Max 71.0925 +trainer/Log Pis Min -9.27914 +trainer/policy/mean Mean -0.0209908 +trainer/policy/mean Std 0.908539 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77893 +trainer/policy/normal/std Std 0.657288 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.336518 +trainer/policy/normal/log_std Mean 0.982027 +trainer/policy/normal/log_std Std 0.319732 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.0891 +eval/num steps total 493385 +eval/num paths total 497 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.091706 +eval/Actions Std 0.911801 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3379 +time/logging (s) 0.00379254 +time/sampling batch (s) 0.279064 +time/saving (s) 0.00357744 +time/training (s) 6.3263 +time/epoch (s) 8.95063 +time/total (s) 5095.04 +Epoch -504 +---------------------------------- --------------- +2022-05-10 14:35:47.069790 PDT | [0] Epoch -503 finished +---------------------------------- --------------- +epoch -503 +replay_buffer/size 999033 +trainer/num train calls 498000 +trainer/Policy Loss -20.3661 +trainer/Log Pis Mean 24.2992 +trainer/Log Pis Std 13.4569 +trainer/Log Pis Max 73.0968 +trainer/Log Pis Min -4.97545 +trainer/policy/mean Mean -0.0426294 +trainer/policy/mean Std 0.903992 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81185 +trainer/policy/normal/std Std 0.657352 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.301651 +trainer/policy/normal/log_std Mean 0.994787 +trainer/policy/normal/log_std Std 0.315405 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19848 +eval/num steps total 494385 +eval/num paths total 498 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.14665 +eval/Actions Std 0.90411 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77695 +time/logging (s) 0.00376131 +time/sampling batch (s) 0.281994 +time/saving (s) 0.00340875 +time/training (s) 6.96533 +time/epoch (s) 10.0314 +time/total (s) 5105.07 +Epoch -503 +---------------------------------- --------------- +2022-05-10 14:35:56.859489 PDT | [0] Epoch -502 finished +---------------------------------- --------------- +epoch -502 +replay_buffer/size 999033 +trainer/num train calls 499000 +trainer/Policy Loss -18.708 +trainer/Log Pis Mean 23.94 +trainer/Log Pis Std 13.0835 +trainer/Log Pis Max 71.7595 +trainer/Log Pis Min -5.47331 +trainer/policy/mean Mean -0.0447447 +trainer/policy/mean Std 0.90391 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79225 +trainer/policy/normal/std Std 0.628344 +trainer/policy/normal/std Max 5.77802 +trainer/policy/normal/std Min 0.256544 +trainer/policy/normal/log_std Mean 0.990807 +trainer/policy/normal/log_std Std 0.302567 +trainer/policy/normal/log_std Max 1.75406 +trainer/policy/normal/log_std Min -1.36046 +eval/num steps total 495385 +eval/num paths total 499 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114501 +eval/Actions Std 0.901096 +eval/Actions Max 0.999998 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47406 +time/logging (s) 0.00370693 +time/sampling batch (s) 0.282261 +time/saving (s) 0.00348869 +time/training (s) 7.0046 +time/epoch (s) 9.76812 +time/total (s) 5114.84 +Epoch -502 +---------------------------------- --------------- +2022-05-10 14:36:06.272706 PDT | [0] Epoch -501 finished +---------------------------------- --------------- +epoch -501 +replay_buffer/size 999033 +trainer/num train calls 500000 +trainer/Policy Loss -20.9633 +trainer/Log Pis Mean 24.1148 +trainer/Log Pis Std 12.8676 +trainer/Log Pis Max 63.1045 +trainer/Log Pis Min -11.0939 +trainer/policy/mean Mean -0.0513611 +trainer/policy/mean Std 0.906303 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79296 +trainer/policy/normal/std Std 0.650303 +trainer/policy/normal/std Max 5.56633 +trainer/policy/normal/std Min 0.300536 +trainer/policy/normal/log_std Mean 0.986134 +trainer/policy/normal/log_std Std 0.327488 +trainer/policy/normal/log_std Max 1.71674 +trainer/policy/normal/log_std Min -1.20219 +eval/num steps total 496385 +eval/num paths total 500 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0866269 +eval/Actions Std 0.915837 +eval/Actions Max 0.999991 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55785 +time/logging (s) 0.00435884 +time/sampling batch (s) 0.280783 +time/saving (s) 0.0072468 +time/training (s) 6.54197 +time/epoch (s) 9.39221 +time/total (s) 5124.24 +Epoch -501 +---------------------------------- --------------- +2022-05-10 14:36:16.100892 PDT | [0] Epoch -500 finished +---------------------------------- --------------- +epoch -500 +replay_buffer/size 999033 +trainer/num train calls 501000 +trainer/Policy Loss -19.15 +trainer/Log Pis Mean 24.4515 +trainer/Log Pis Std 12.3173 +trainer/Log Pis Max 55.597 +trainer/Log Pis Min -6.91705 +trainer/policy/mean Mean -0.0185856 +trainer/policy/mean Std 0.903461 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79557 +trainer/policy/normal/std Std 0.642127 +trainer/policy/normal/std Max 5.502 +trainer/policy/normal/std Min 0.327008 +trainer/policy/normal/log_std Mean 0.98915 +trainer/policy/normal/log_std Std 0.316023 +trainer/policy/normal/log_std Max 1.70511 +trainer/policy/normal/log_std Min -1.11777 +eval/num steps total 497385 +eval/num paths total 501 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105484 +eval/Actions Std 0.929199 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53945 +time/logging (s) 0.00378364 +time/sampling batch (s) 0.281196 +time/saving (s) 0.00365032 +time/training (s) 6.97731 +time/epoch (s) 9.80539 +time/total (s) 5134.05 +Epoch -500 +---------------------------------- --------------- +2022-05-10 14:36:26.145328 PDT | [0] Epoch -499 finished +---------------------------------- --------------- +epoch -499 +replay_buffer/size 999033 +trainer/num train calls 502000 +trainer/Policy Loss -19.2691 +trainer/Log Pis Mean 24.0676 +trainer/Log Pis Std 13.3315 +trainer/Log Pis Max 68.6737 +trainer/Log Pis Min -7.36577 +trainer/policy/mean Mean -0.0444182 +trainer/policy/mean Std 0.908307 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.805 +trainer/policy/normal/std Std 0.653776 +trainer/policy/normal/std Max 6.16566 +trainer/policy/normal/std Min 0.247378 +trainer/policy/normal/log_std Mean 0.991494 +trainer/policy/normal/log_std Std 0.321329 +trainer/policy/normal/log_std Max 1.81899 +trainer/policy/normal/log_std Min -1.39684 +eval/num steps total 498385 +eval/num paths total 502 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0126647 +eval/Actions Std 0.945197 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41926 +time/logging (s) 0.00364266 +time/sampling batch (s) 0.280267 +time/saving (s) 0.00342327 +time/training (s) 7.31571 +time/epoch (s) 10.0223 +time/total (s) 5144.07 +Epoch -499 +---------------------------------- --------------- +2022-05-10 14:36:37.042036 PDT | [0] Epoch -498 finished +---------------------------------- --------------- +epoch -498 +replay_buffer/size 999033 +trainer/num train calls 503000 +trainer/Policy Loss -19.496 +trainer/Log Pis Mean 25.2016 +trainer/Log Pis Std 13.474 +trainer/Log Pis Max 83.4027 +trainer/Log Pis Min -9.64835 +trainer/policy/mean Mean -0.0526295 +trainer/policy/mean Std 0.906846 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76492 +trainer/policy/normal/std Std 0.626494 +trainer/policy/normal/std Max 6.30837 +trainer/policy/normal/std Min 0.292948 +trainer/policy/normal/log_std Mean 0.980711 +trainer/policy/normal/log_std Std 0.302466 +trainer/policy/normal/log_std Max 1.84188 +trainer/policy/normal/log_std Min -1.22776 +eval/num steps total 499385 +eval/num paths total 503 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108848 +eval/Actions Std 0.881582 +eval/Actions Max 0.999991 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58033 +time/logging (s) 0.00374266 +time/sampling batch (s) 0.280936 +time/saving (s) 0.00349376 +time/training (s) 8.00648 +time/epoch (s) 10.875 +time/total (s) 5154.95 +Epoch -498 +---------------------------------- --------------- +2022-05-10 14:36:47.115505 PDT | [0] Epoch -497 finished +---------------------------------- --------------- +epoch -497 +replay_buffer/size 999033 +trainer/num train calls 504000 +trainer/Policy Loss -20.1742 +trainer/Log Pis Mean 25.625 +trainer/Log Pis Std 13.3582 +trainer/Log Pis Max 65.3981 +trainer/Log Pis Min -4.56662 +trainer/policy/mean Mean -0.0473917 +trainer/policy/mean Std 0.911033 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79562 +trainer/policy/normal/std Std 0.640572 +trainer/policy/normal/std Max 7.05242 +trainer/policy/normal/std Min 0.326654 +trainer/policy/normal/log_std Mean 0.990069 +trainer/policy/normal/log_std Std 0.310825 +trainer/policy/normal/log_std Max 1.95337 +trainer/policy/normal/log_std Min -1.11885 +eval/num steps total 500385 +eval/num paths total 504 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.111374 +eval/Actions Std 0.906121 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65504 +time/logging (s) 0.00378507 +time/sampling batch (s) 0.283577 +time/saving (s) 0.00352566 +time/training (s) 7.10596 +time/epoch (s) 10.0519 +time/total (s) 5165 +Epoch -497 +---------------------------------- --------------- +2022-05-10 14:36:57.487499 PDT | [0] Epoch -496 finished +---------------------------------- --------------- +epoch -496 +replay_buffer/size 999033 +trainer/num train calls 505000 +trainer/Policy Loss -18.9376 +trainer/Log Pis Mean 24.5569 +trainer/Log Pis Std 13.5193 +trainer/Log Pis Max 66.2124 +trainer/Log Pis Min -6.72808 +trainer/policy/mean Mean -0.0599495 +trainer/policy/mean Std 0.905058 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77179 +trainer/policy/normal/std Std 0.619417 +trainer/policy/normal/std Max 5.05605 +trainer/policy/normal/std Min 0.327446 +trainer/policy/normal/log_std Mean 0.983597 +trainer/policy/normal/log_std Std 0.30185 +trainer/policy/normal/log_std Max 1.62059 +trainer/policy/normal/log_std Min -1.11643 +eval/num steps total 501385 +eval/num paths total 505 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0983701 +eval/Actions Std 0.918019 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68822 +time/logging (s) 0.00371559 +time/sampling batch (s) 0.535555 +time/saving (s) 0.00346273 +time/training (s) 7.11919 +time/epoch (s) 10.3501 +time/total (s) 5175.36 +Epoch -496 +---------------------------------- --------------- +2022-05-10 14:37:06.848560 PDT | [0] Epoch -495 finished +---------------------------------- --------------- +epoch -495 +replay_buffer/size 999033 +trainer/num train calls 506000 +trainer/Policy Loss -21.731 +trainer/Log Pis Mean 24.3078 +trainer/Log Pis Std 13.1532 +trainer/Log Pis Max 67.6031 +trainer/Log Pis Min -11.2289 +trainer/policy/mean Mean -0.0388384 +trainer/policy/mean Std 0.906587 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.89265 +trainer/policy/normal/std Std 0.643746 +trainer/policy/normal/std Max 5.90994 +trainer/policy/normal/std Min 0.37758 +trainer/policy/normal/log_std Mean 1.02792 +trainer/policy/normal/log_std Std 0.29104 +trainer/policy/normal/log_std Max 1.77664 +trainer/policy/normal/log_std Min -0.973974 +eval/num steps total 502385 +eval/num paths total 506 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138303 +eval/Actions Std 0.89116 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67859 +time/logging (s) 0.00400692 +time/sampling batch (s) 0.279016 +time/saving (s) 0.00405915 +time/training (s) 6.3742 +time/epoch (s) 9.33988 +time/total (s) 5184.7 +Epoch -495 +---------------------------------- --------------- +2022-05-10 14:37:16.404625 PDT | [0] Epoch -494 finished +---------------------------------- --------------- +epoch -494 +replay_buffer/size 999033 +trainer/num train calls 507000 +trainer/Policy Loss -18.6479 +trainer/Log Pis Mean 24.2133 +trainer/Log Pis Std 13.2421 +trainer/Log Pis Max 69.373 +trainer/Log Pis Min -7.97982 +trainer/policy/mean Mean -0.0479722 +trainer/policy/mean Std 0.906446 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.81804 +trainer/policy/normal/std Std 0.610276 +trainer/policy/normal/std Max 6.06681 +trainer/policy/normal/std Min 0.361217 +trainer/policy/normal/log_std Mean 1.00386 +trainer/policy/normal/log_std Std 0.281547 +trainer/policy/normal/log_std Max 1.80283 +trainer/policy/normal/log_std Min -1.01828 +eval/num steps total 503385 +eval/num paths total 507 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132867 +eval/Actions Std 0.910441 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54464 +time/logging (s) 0.00378609 +time/sampling batch (s) 0.278914 +time/saving (s) 0.00352899 +time/training (s) 6.70341 +time/epoch (s) 9.53427 +time/total (s) 5194.24 +Epoch -494 +---------------------------------- --------------- +2022-05-10 14:37:25.768818 PDT | [0] Epoch -493 finished +---------------------------------- --------------- +epoch -493 +replay_buffer/size 999033 +trainer/num train calls 508000 +trainer/Policy Loss -19.596 +trainer/Log Pis Mean 24.1026 +trainer/Log Pis Std 12.7792 +trainer/Log Pis Max 62.3354 +trainer/Log Pis Min -7.23615 +trainer/policy/mean Mean -0.0551187 +trainer/policy/mean Std 0.903825 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81241 +trainer/policy/normal/std Std 0.659534 +trainer/policy/normal/std Max 5.5399 +trainer/policy/normal/std Min 0.320457 +trainer/policy/normal/log_std Mean 0.993706 +trainer/policy/normal/log_std Std 0.321475 +trainer/policy/normal/log_std Max 1.71198 +trainer/policy/normal/log_std Min -1.13801 +eval/num steps total 504385 +eval/num paths total 508 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0924781 +eval/Actions Std 0.890565 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69174 +time/logging (s) 0.00369842 +time/sampling batch (s) 0.277049 +time/saving (s) 0.00345102 +time/training (s) 6.36701 +time/epoch (s) 9.34295 +time/total (s) 5203.58 +Epoch -493 +---------------------------------- --------------- +2022-05-10 14:37:35.579932 PDT | [0] Epoch -492 finished +---------------------------------- --------------- +epoch -492 +replay_buffer/size 999033 +trainer/num train calls 509000 +trainer/Policy Loss -19.8942 +trainer/Log Pis Mean 24.3477 +trainer/Log Pis Std 13.6834 +trainer/Log Pis Max 78.975 +trainer/Log Pis Min -8.34395 +trainer/policy/mean Mean -0.0255843 +trainer/policy/mean Std 0.902836 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82034 +trainer/policy/normal/std Std 0.659656 +trainer/policy/normal/std Max 5.6164 +trainer/policy/normal/std Min 0.326415 +trainer/policy/normal/log_std Mean 0.99784 +trainer/policy/normal/log_std Std 0.313918 +trainer/policy/normal/log_std Max 1.72569 +trainer/policy/normal/log_std Min -1.11958 +eval/num steps total 505385 +eval/num paths total 509 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121609 +eval/Actions Std 0.905543 +eval/Actions Max 0.99999 +eval/Actions Min -0.999973 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68508 +time/logging (s) 0.00378809 +time/sampling batch (s) 0.277175 +time/saving (s) 0.00346271 +time/training (s) 6.82047 +time/epoch (s) 9.78998 +time/total (s) 5213.38 +Epoch -492 +---------------------------------- --------------- +2022-05-10 14:37:45.905210 PDT | [0] Epoch -491 finished +---------------------------------- --------------- +epoch -491 +replay_buffer/size 999033 +trainer/num train calls 510000 +trainer/Policy Loss -19.4065 +trainer/Log Pis Mean 23.8056 +trainer/Log Pis Std 12.6987 +trainer/Log Pis Max 63.3985 +trainer/Log Pis Min -7.89934 +trainer/policy/mean Mean -0.0452828 +trainer/policy/mean Std 0.903869 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7516 +trainer/policy/normal/std Std 0.641282 +trainer/policy/normal/std Max 6.16317 +trainer/policy/normal/std Min 0.294488 +trainer/policy/normal/log_std Mean 0.971961 +trainer/policy/normal/log_std Std 0.322299 +trainer/policy/normal/log_std Max 1.81859 +trainer/policy/normal/log_std Min -1.22252 +eval/num steps total 506385 +eval/num paths total 510 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0356476 +eval/Actions Std 0.927616 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48286 +time/logging (s) 0.00374835 +time/sampling batch (s) 0.276869 +time/saving (s) 0.00348825 +time/training (s) 7.53712 +time/epoch (s) 10.3041 +time/total (s) 5223.68 +Epoch -491 +---------------------------------- --------------- +2022-05-10 14:37:55.946170 PDT | [0] Epoch -490 finished +---------------------------------- --------------- +epoch -490 +replay_buffer/size 999033 +trainer/num train calls 511000 +trainer/Policy Loss -20.0799 +trainer/Log Pis Mean 24.3817 +trainer/Log Pis Std 13.4795 +trainer/Log Pis Max 67.3301 +trainer/Log Pis Min -12.6297 +trainer/policy/mean Mean -0.0337438 +trainer/policy/mean Std 0.907974 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8458 +trainer/policy/normal/std Std 0.662093 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.29372 +trainer/policy/normal/log_std Mean 1.00803 +trainer/policy/normal/log_std Std 0.308277 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22513 +eval/num steps total 507385 +eval/num paths total 511 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.416876 +eval/Actions Std 0.705821 +eval/Actions Max 0.999997 +eval/Actions Min -0.99996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41235 +time/logging (s) 0.00378905 +time/sampling batch (s) 0.277443 +time/saving (s) 0.0034449 +time/training (s) 7.32278 +time/epoch (s) 10.0198 +time/total (s) 5233.7 +Epoch -490 +---------------------------------- --------------- +2022-05-10 14:38:05.458384 PDT | [0] Epoch -489 finished +---------------------------------- --------------- +epoch -489 +replay_buffer/size 999033 +trainer/num train calls 512000 +trainer/Policy Loss -18.9814 +trainer/Log Pis Mean 25.5677 +trainer/Log Pis Std 13.7793 +trainer/Log Pis Max 81.5892 +trainer/Log Pis Min -7.28098 +trainer/policy/mean Mean -0.0380244 +trainer/policy/mean Std 0.905172 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77276 +trainer/policy/normal/std Std 0.632781 +trainer/policy/normal/std Max 5.82551 +trainer/policy/normal/std Min 0.31281 +trainer/policy/normal/log_std Mean 0.981083 +trainer/policy/normal/log_std Std 0.316988 +trainer/policy/normal/log_std Max 1.76225 +trainer/policy/normal/log_std Min -1.16216 +eval/num steps total 508385 +eval/num paths total 512 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112571 +eval/Actions Std 0.913249 +eval/Actions Max 0.999988 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65337 +time/logging (s) 0.00405029 +time/sampling batch (s) 0.276816 +time/saving (s) 0.00412031 +time/training (s) 6.55281 +time/epoch (s) 9.49117 +time/total (s) 5243.2 +Epoch -489 +---------------------------------- --------------- +2022-05-10 14:38:15.543638 PDT | [0] Epoch -488 finished +---------------------------------- --------------- +epoch -488 +replay_buffer/size 999033 +trainer/num train calls 513000 +trainer/Policy Loss -19.4289 +trainer/Log Pis Mean 25.1237 +trainer/Log Pis Std 13.6466 +trainer/Log Pis Max 82.7757 +trainer/Log Pis Min -10.1453 +trainer/policy/mean Mean -0.0556339 +trainer/policy/mean Std 0.904879 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78088 +trainer/policy/normal/std Std 0.654524 +trainer/policy/normal/std Max 6.44978 +trainer/policy/normal/std Min 0.287624 +trainer/policy/normal/log_std Mean 0.983336 +trainer/policy/normal/log_std Std 0.316202 +trainer/policy/normal/log_std Max 1.86405 +trainer/policy/normal/log_std Min -1.2461 +eval/num steps total 509385 +eval/num paths total 513 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105837 +eval/Actions Std 0.919776 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54749 +time/logging (s) 0.00365729 +time/sampling batch (s) 0.278498 +time/saving (s) 0.0033745 +time/training (s) 7.23021 +time/epoch (s) 10.0632 +time/total (s) 5253.27 +Epoch -488 +---------------------------------- --------------- +2022-05-10 14:38:25.780781 PDT | [0] Epoch -487 finished +---------------------------------- --------------- +epoch -487 +replay_buffer/size 999033 +trainer/num train calls 514000 +trainer/Policy Loss -21.0404 +trainer/Log Pis Mean 24.9074 +trainer/Log Pis Std 13.4434 +trainer/Log Pis Max 64.8022 +trainer/Log Pis Min -5.24464 +trainer/policy/mean Mean -0.0449491 +trainer/policy/mean Std 0.909195 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84668 +trainer/policy/normal/std Std 0.617109 +trainer/policy/normal/std Max 5.58066 +trainer/policy/normal/std Min 0.364556 +trainer/policy/normal/log_std Mean 1.01284 +trainer/policy/normal/log_std Std 0.290213 +trainer/policy/normal/log_std Max 1.71931 +trainer/policy/normal/log_std Min -1.00908 +eval/num steps total 510385 +eval/num paths total 514 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0199429 +eval/Actions Std 0.907931 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56081 +time/logging (s) 0.00375278 +time/sampling batch (s) 0.277392 +time/saving (s) 0.00341671 +time/training (s) 7.37063 +time/epoch (s) 10.216 +time/total (s) 5263.48 +Epoch -487 +---------------------------------- --------------- +2022-05-10 14:38:36.453831 PDT | [0] Epoch -486 finished +---------------------------------- --------------- +epoch -486 +replay_buffer/size 999033 +trainer/num train calls 515000 +trainer/Policy Loss -20.2401 +trainer/Log Pis Mean 24.2324 +trainer/Log Pis Std 13.3147 +trainer/Log Pis Max 76.2202 +trainer/Log Pis Min -11.8742 +trainer/policy/mean Mean -0.0373903 +trainer/policy/mean Std 0.902263 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.75074 +trainer/policy/normal/std Std 0.657769 +trainer/policy/normal/std Max 5.91902 +trainer/policy/normal/std Min 0.183939 +trainer/policy/normal/log_std Mean 0.969261 +trainer/policy/normal/log_std Std 0.333099 +trainer/policy/normal/log_std Max 1.77817 +trainer/policy/normal/log_std Min -1.69315 +eval/num steps total 511385 +eval/num paths total 515 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.212552 +eval/Actions Std 0.962967 +eval/Actions Max 0.999989 +eval/Actions Min -0.999977 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61337 +time/logging (s) 0.00372576 +time/sampling batch (s) 0.281036 +time/saving (s) 0.00344659 +time/training (s) 7.74973 +time/epoch (s) 10.6513 +time/total (s) 5274.14 +Epoch -486 +---------------------------------- --------------- +2022-05-10 14:38:47.242643 PDT | [0] Epoch -485 finished +---------------------------------- --------------- +epoch -485 +replay_buffer/size 999033 +trainer/num train calls 516000 +trainer/Policy Loss -19.8571 +trainer/Log Pis Mean 25.0408 +trainer/Log Pis Std 13.685 +trainer/Log Pis Max 67.5401 +trainer/Log Pis Min -9.28504 +trainer/policy/mean Mean -0.0253146 +trainer/policy/mean Std 0.910043 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82536 +trainer/policy/normal/std Std 0.641927 +trainer/policy/normal/std Max 6.45916 +trainer/policy/normal/std Min 0.315541 +trainer/policy/normal/log_std Mean 1.00232 +trainer/policy/normal/log_std Std 0.302665 +trainer/policy/normal/log_std Max 1.8655 +trainer/policy/normal/log_std Min -1.15346 +eval/num steps total 512385 +eval/num paths total 516 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00871599 +eval/Actions Std 0.969654 +eval/Actions Max 0.999995 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56581 +time/logging (s) 0.0041343 +time/sampling batch (s) 0.282955 +time/saving (s) 0.00418351 +time/training (s) 7.65989 +time/epoch (s) 10.517 +time/total (s) 5284.91 +Epoch -485 +---------------------------------- --------------- +2022-05-10 14:38:58.778206 PDT | [0] Epoch -484 finished +---------------------------------- --------------- +epoch -484 +replay_buffer/size 999033 +trainer/num train calls 517000 +trainer/Policy Loss -20.7486 +trainer/Log Pis Mean 25.0646 +trainer/Log Pis Std 13.9784 +trainer/Log Pis Max 71.8561 +trainer/Log Pis Min -12.4097 +trainer/policy/mean Mean -0.0245137 +trainer/policy/mean Std 0.901946 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.82742 +trainer/policy/normal/std Std 0.65546 +trainer/policy/normal/std Max 5.59786 +trainer/policy/normal/std Min 0.310795 +trainer/policy/normal/log_std Mean 1.00004 +trainer/policy/normal/log_std Std 0.318026 +trainer/policy/normal/log_std Max 1.72238 +trainer/policy/normal/log_std Min -1.16862 +eval/num steps total 513385 +eval/num paths total 517 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.182207 +eval/Actions Std 0.94849 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67952 +time/logging (s) 0.0037044 +time/sampling batch (s) 0.283837 +time/saving (s) 0.00355122 +time/training (s) 8.54233 +time/epoch (s) 11.5129 +time/total (s) 5296.42 +Epoch -484 +---------------------------------- --------------- +2022-05-10 14:39:09.391371 PDT | [0] Epoch -483 finished +---------------------------------- --------------- +epoch -483 +replay_buffer/size 999033 +trainer/num train calls 518000 +trainer/Policy Loss -18.8534 +trainer/Log Pis Mean 23.7468 +trainer/Log Pis Std 13.3233 +trainer/Log Pis Max 69.5845 +trainer/Log Pis Min -7.44423 +trainer/policy/mean Mean 0.00591347 +trainer/policy/mean Std 0.900988 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82715 +trainer/policy/normal/std Std 0.657829 +trainer/policy/normal/std Max 6.08524 +trainer/policy/normal/std Min 0.356476 +trainer/policy/normal/log_std Mean 0.999671 +trainer/policy/normal/log_std Std 0.319106 +trainer/policy/normal/log_std Max 1.80587 +trainer/policy/normal/log_std Min -1.03149 +eval/num steps total 514385 +eval/num paths total 518 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.239108 +eval/Actions Std 0.879591 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61947 +time/logging (s) 0.00366339 +time/sampling batch (s) 0.280904 +time/saving (s) 0.00349644 +time/training (s) 7.68389 +time/epoch (s) 10.5914 +time/total (s) 5307.02 +Epoch -483 +---------------------------------- --------------- +2022-05-10 14:39:20.827365 PDT | [0] Epoch -482 finished +---------------------------------- --------------- +epoch -482 +replay_buffer/size 999033 +trainer/num train calls 519000 +trainer/Policy Loss -19.5129 +trainer/Log Pis Mean 24.9258 +trainer/Log Pis Std 13.5072 +trainer/Log Pis Max 72.5277 +trainer/Log Pis Min -6.53332 +trainer/policy/mean Mean -0.031247 +trainer/policy/mean Std 0.90815 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85366 +trainer/policy/normal/std Std 0.644091 +trainer/policy/normal/std Max 5.81912 +trainer/policy/normal/std Min 0.306902 +trainer/policy/normal/log_std Mean 1.01101 +trainer/policy/normal/log_std Std 0.311789 +trainer/policy/normal/log_std Max 1.76115 +trainer/policy/normal/log_std Min -1.18123 +eval/num steps total 515385 +eval/num paths total 519 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0279395 +eval/Actions Std 0.907311 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7039 +time/logging (s) 0.0041352 +time/sampling batch (s) 0.544722 +time/saving (s) 0.00416153 +time/training (s) 8.15713 +time/epoch (s) 11.414 +time/total (s) 5318.44 +Epoch -482 +---------------------------------- --------------- +2022-05-10 14:39:31.100850 PDT | [0] Epoch -481 finished +---------------------------------- --------------- +epoch -481 +replay_buffer/size 999033 +trainer/num train calls 520000 +trainer/Policy Loss -19.3134 +trainer/Log Pis Mean 25.2619 +trainer/Log Pis Std 13.7506 +trainer/Log Pis Max 69.1128 +trainer/Log Pis Min -11.4135 +trainer/policy/mean Mean -0.0252849 +trainer/policy/mean Std 0.91026 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80752 +trainer/policy/normal/std Std 0.67717 +trainer/policy/normal/std Max 7.35534 +trainer/policy/normal/std Min 0.250737 +trainer/policy/normal/log_std Mean 0.988379 +trainer/policy/normal/log_std Std 0.338845 +trainer/policy/normal/log_std Max 1.99543 +trainer/policy/normal/log_std Min -1.38335 +eval/num steps total 516385 +eval/num paths total 520 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0679675 +eval/Actions Std 0.917955 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51219 +time/logging (s) 0.00370168 +time/sampling batch (s) 0.282653 +time/saving (s) 0.00356518 +time/training (s) 7.44898 +time/epoch (s) 10.2511 +time/total (s) 5328.69 +Epoch -481 +---------------------------------- --------------- +2022-05-10 14:39:41.177585 PDT | [0] Epoch -480 finished +---------------------------------- --------------- +epoch -480 +replay_buffer/size 999033 +trainer/num train calls 521000 +trainer/Policy Loss -19.5503 +trainer/Log Pis Mean 24.8989 +trainer/Log Pis Std 13.4807 +trainer/Log Pis Max 67.8992 +trainer/Log Pis Min -6.98366 +trainer/policy/mean Mean -0.039633 +trainer/policy/mean Std 0.907648 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8641 +trainer/policy/normal/std Std 0.663473 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.294634 +trainer/policy/normal/log_std Mean 1.01304 +trainer/policy/normal/log_std Std 0.317854 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22202 +eval/num steps total 517385 +eval/num paths total 521 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118432 +eval/Actions Std 0.898824 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58476 +time/logging (s) 0.00368736 +time/sampling batch (s) 0.530857 +time/saving (s) 0.00342912 +time/training (s) 6.93242 +time/epoch (s) 10.0552 +time/total (s) 5338.75 +Epoch -480 +---------------------------------- --------------- +2022-05-10 14:39:52.101527 PDT | [0] Epoch -479 finished +---------------------------------- --------------- +epoch -479 +replay_buffer/size 999033 +trainer/num train calls 522000 +trainer/Policy Loss -18.8954 +trainer/Log Pis Mean 24.419 +trainer/Log Pis Std 12.9532 +trainer/Log Pis Max 67.156 +trainer/Log Pis Min -10.1595 +trainer/policy/mean Mean -0.0391311 +trainer/policy/mean Std 0.904892 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80606 +trainer/policy/normal/std Std 0.646874 +trainer/policy/normal/std Max 5.4211 +trainer/policy/normal/std Min 0.292592 +trainer/policy/normal/log_std Mean 0.992971 +trainer/policy/normal/log_std Std 0.315625 +trainer/policy/normal/log_std Max 1.6903 +trainer/policy/normal/log_std Min -1.22898 +eval/num steps total 518385 +eval/num paths total 522 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.325852 +eval/Actions Std 0.801442 +eval/Actions Max 0.999997 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56177 +time/logging (s) 0.00376696 +time/sampling batch (s) 0.781847 +time/saving (s) 0.00345633 +time/training (s) 7.55167 +time/epoch (s) 10.9025 +time/total (s) 5349.65 +Epoch -479 +---------------------------------- --------------- +2022-05-10 14:40:01.470537 PDT | [0] Epoch -478 finished +---------------------------------- --------------- +epoch -478 +replay_buffer/size 999033 +trainer/num train calls 523000 +trainer/Policy Loss -19.8977 +trainer/Log Pis Mean 25.1813 +trainer/Log Pis Std 14.1363 +trainer/Log Pis Max 72.2282 +trainer/Log Pis Min -7.4879 +trainer/policy/mean Mean -0.0565525 +trainer/policy/mean Std 0.905904 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.85178 +trainer/policy/normal/std Std 0.638778 +trainer/policy/normal/std Max 6.41313 +trainer/policy/normal/std Min 0.289093 +trainer/policy/normal/log_std Mean 1.01274 +trainer/policy/normal/log_std Std 0.298292 +trainer/policy/normal/log_std Max 1.85835 +trainer/policy/normal/log_std Min -1.24101 +eval/num steps total 519385 +eval/num paths total 523 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.198824 +eval/Actions Std 0.952445 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70561 +time/logging (s) 0.00414348 +time/sampling batch (s) 0.281622 +time/saving (s) 0.00420272 +time/training (s) 6.35214 +time/epoch (s) 9.34771 +time/total (s) 5359 +Epoch -478 +---------------------------------- --------------- +2022-05-10 14:40:12.717955 PDT | [0] Epoch -477 finished +---------------------------------- --------------- +epoch -477 +replay_buffer/size 999033 +trainer/num train calls 524000 +trainer/Policy Loss -20.9779 +trainer/Log Pis Mean 24.0541 +trainer/Log Pis Std 13.7453 +trainer/Log Pis Max 76.0139 +trainer/Log Pis Min -7.55943 +trainer/policy/mean Mean -0.0372647 +trainer/policy/mean Std 0.904588 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7527 +trainer/policy/normal/std Std 0.644004 +trainer/policy/normal/std Max 6.16478 +trainer/policy/normal/std Min 0.27257 +trainer/policy/normal/log_std Mean 0.972271 +trainer/policy/normal/log_std Std 0.322906 +trainer/policy/normal/log_std Max 1.81885 +trainer/policy/normal/log_std Min -1.29986 +eval/num steps total 520385 +eval/num paths total 524 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0618888 +eval/Actions Std 0.895212 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66589 +time/logging (s) 0.00659978 +time/sampling batch (s) 0.781594 +time/saving (s) 0.00528142 +time/training (s) 7.76851 +time/epoch (s) 11.2279 +time/total (s) 5370.24 +Epoch -477 +---------------------------------- --------------- +2022-05-10 14:40:23.174328 PDT | [0] Epoch -476 finished +---------------------------------- --------------- +epoch -476 +replay_buffer/size 999033 +trainer/num train calls 525000 +trainer/Policy Loss -19.7823 +trainer/Log Pis Mean 25.6897 +trainer/Log Pis Std 13.3307 +trainer/Log Pis Max 81.5057 +trainer/Log Pis Min -5.6635 +trainer/policy/mean Mean -0.058394 +trainer/policy/mean Std 0.90524 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.88554 +trainer/policy/normal/std Std 0.687566 +trainer/policy/normal/std Max 6.45743 +trainer/policy/normal/std Min 0.3257 +trainer/policy/normal/log_std Mean 1.0181 +trainer/policy/normal/log_std Std 0.327241 +trainer/policy/normal/log_std Max 1.86523 +trainer/policy/normal/log_std Min -1.12178 +eval/num steps total 521385 +eval/num paths total 525 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.180322 +eval/Actions Std 0.911243 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7008 +time/logging (s) 0.00388264 +time/sampling batch (s) 0.282708 +time/saving (s) 0.00341565 +time/training (s) 7.43904 +time/epoch (s) 10.4298 +time/total (s) 5380.67 +Epoch -476 +---------------------------------- --------------- +2022-05-10 14:40:34.751561 PDT | [0] Epoch -475 finished +---------------------------------- --------------- +epoch -475 +replay_buffer/size 999033 +trainer/num train calls 526000 +trainer/Policy Loss -19.5344 +trainer/Log Pis Mean 24.0142 +trainer/Log Pis Std 13.5177 +trainer/Log Pis Max 66.186 +trainer/Log Pis Min -10.7121 +trainer/policy/mean Mean -0.0247106 +trainer/policy/mean Std 0.906835 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81059 +trainer/policy/normal/std Std 0.671415 +trainer/policy/normal/std Max 6.40219 +trainer/policy/normal/std Min 0.313294 +trainer/policy/normal/log_std Mean 0.990801 +trainer/policy/normal/log_std Std 0.332877 +trainer/policy/normal/log_std Max 1.85664 +trainer/policy/normal/log_std Min -1.16061 +eval/num steps total 522385 +eval/num paths total 526 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.157576 +eval/Actions Std 0.88564 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85038 +time/logging (s) 0.00397409 +time/sampling batch (s) 1.0339 +time/saving (s) 0.00347968 +time/training (s) 7.66381 +time/epoch (s) 11.5555 +time/total (s) 5392.23 +Epoch -475 +---------------------------------- --------------- +2022-05-10 14:40:45.430174 PDT | [0] Epoch -474 finished +---------------------------------- --------------- +epoch -474 +replay_buffer/size 999033 +trainer/num train calls 527000 +trainer/Policy Loss -19.0291 +trainer/Log Pis Mean 23.3922 +trainer/Log Pis Std 13.8544 +trainer/Log Pis Max 80.3258 +trainer/Log Pis Min -9.59969 +trainer/policy/mean Mean -0.0224419 +trainer/policy/mean Std 0.901603 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.79978 +trainer/policy/normal/std Std 0.647452 +trainer/policy/normal/std Max 5.03031 +trainer/policy/normal/std Min 0.26232 +trainer/policy/normal/log_std Mean 0.990954 +trainer/policy/normal/log_std Std 0.312943 +trainer/policy/normal/log_std Max 1.61548 +trainer/policy/normal/log_std Min -1.33819 +eval/num steps total 523385 +eval/num paths total 527 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0592428 +eval/Actions Std 0.915386 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60776 +time/logging (s) 0.00378252 +time/sampling batch (s) 0.280141 +time/saving (s) 0.00345875 +time/training (s) 7.76159 +time/epoch (s) 10.6567 +time/total (s) 5402.89 +Epoch -474 +---------------------------------- --------------- +2022-05-10 14:40:56.732948 PDT | [0] Epoch -473 finished +---------------------------------- --------------- +epoch -473 +replay_buffer/size 999033 +trainer/num train calls 528000 +trainer/Policy Loss -19.3941 +trainer/Log Pis Mean 24.2248 +trainer/Log Pis Std 12.8705 +trainer/Log Pis Max 67.9931 +trainer/Log Pis Min -8.98468 +trainer/policy/mean Mean -0.0492585 +trainer/policy/mean Std 0.906197 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83486 +trainer/policy/normal/std Std 0.645855 +trainer/policy/normal/std Max 5.87746 +trainer/policy/normal/std Min 0.269954 +trainer/policy/normal/log_std Mean 1.00507 +trainer/policy/normal/log_std Std 0.305462 +trainer/policy/normal/log_std Max 1.77113 +trainer/policy/normal/log_std Min -1.3095 +eval/num steps total 524385 +eval/num paths total 528 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126499 +eval/Actions Std 0.905262 +eval/Actions Max 0.999989 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68372 +time/logging (s) 0.00416024 +time/sampling batch (s) 0.534047 +time/saving (s) 0.00403767 +time/training (s) 8.05519 +time/epoch (s) 11.2812 +time/total (s) 5414.17 +Epoch -473 +---------------------------------- --------------- +2022-05-10 14:41:06.715368 PDT | [0] Epoch -472 finished +---------------------------------- --------------- +epoch -472 +replay_buffer/size 999033 +trainer/num train calls 529000 +trainer/Policy Loss -19.9839 +trainer/Log Pis Mean 24.7446 +trainer/Log Pis Std 13.2507 +trainer/Log Pis Max 68.0498 +trainer/Log Pis Min -4.25224 +trainer/policy/mean Mean -0.0356496 +trainer/policy/mean Std 0.908847 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.81081 +trainer/policy/normal/std Std 0.624139 +trainer/policy/normal/std Max 5.23986 +trainer/policy/normal/std Min 0.338329 +trainer/policy/normal/log_std Mean 0.998491 +trainer/policy/normal/log_std Std 0.297032 +trainer/policy/normal/log_std Max 1.6563 +trainer/policy/normal/log_std Min -1.08374 +eval/num steps total 525385 +eval/num paths total 529 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.079405 +eval/Actions Std 0.907176 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67524 +time/logging (s) 0.00380876 +time/sampling batch (s) 0.280678 +time/saving (s) 0.00367614 +time/training (s) 6.9967 +time/epoch (s) 9.9601 +time/total (s) 5424.13 +Epoch -472 +---------------------------------- --------------- +2022-05-10 14:41:16.419513 PDT | [0] Epoch -471 finished +---------------------------------- --------------- +epoch -471 +replay_buffer/size 999033 +trainer/num train calls 530000 +trainer/Policy Loss -20.4309 +trainer/Log Pis Mean 24.8912 +trainer/Log Pis Std 13.8999 +trainer/Log Pis Max 71.4164 +trainer/Log Pis Min -10.9797 +trainer/policy/mean Mean -0.0388996 +trainer/policy/mean Std 0.911824 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84727 +trainer/policy/normal/std Std 0.613867 +trainer/policy/normal/std Max 5.51745 +trainer/policy/normal/std Min 0.297964 +trainer/policy/normal/log_std Mean 1.01427 +trainer/policy/normal/log_std Std 0.283121 +trainer/policy/normal/log_std Max 1.70792 +trainer/policy/normal/log_std Min -1.21078 +eval/num steps total 526385 +eval/num paths total 530 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00652462 +eval/Actions Std 0.987173 +eval/Actions Max 0.999986 +eval/Actions Min -0.999962 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52606 +time/logging (s) 0.00387335 +time/sampling batch (s) 0.530121 +time/saving (s) 0.00345561 +time/training (s) 6.61916 +time/epoch (s) 9.68267 +time/total (s) 5433.82 +Epoch -471 +---------------------------------- --------------- +2022-05-10 14:41:26.407313 PDT | [0] Epoch -470 finished +---------------------------------- --------------- +epoch -470 +replay_buffer/size 999033 +trainer/num train calls 531000 +trainer/Policy Loss -20.0857 +trainer/Log Pis Mean 24.3557 +trainer/Log Pis Std 13.2403 +trainer/Log Pis Max 65.8825 +trainer/Log Pis Min -12.0835 +trainer/policy/mean Mean -0.0165558 +trainer/policy/mean Std 0.910018 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.84907 +trainer/policy/normal/std Std 0.637121 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.286551 +trainer/policy/normal/log_std Mean 1.01233 +trainer/policy/normal/log_std Std 0.294265 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24984 +eval/num steps total 527385 +eval/num paths total 531 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0826724 +eval/Actions Std 0.906318 +eval/Actions Max 0.999992 +eval/Actions Min -0.999981 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65531 +time/logging (s) 0.00370703 +time/sampling batch (s) 0.280858 +time/saving (s) 0.00347063 +time/training (s) 7.02264 +time/epoch (s) 9.96598 +time/total (s) 5443.79 +Epoch -470 +---------------------------------- --------------- +2022-05-10 14:41:36.357218 PDT | [0] Epoch -469 finished +---------------------------------- --------------- +epoch -469 +replay_buffer/size 999033 +trainer/num train calls 532000 +trainer/Policy Loss -20.1122 +trainer/Log Pis Mean 25.157 +trainer/Log Pis Std 12.4724 +trainer/Log Pis Max 68.2537 +trainer/Log Pis Min -4.1466 +trainer/policy/mean Mean -0.0363558 +trainer/policy/mean Std 0.906003 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.8044 +trainer/policy/normal/std Std 0.668309 +trainer/policy/normal/std Max 6.11492 +trainer/policy/normal/std Min 0.280818 +trainer/policy/normal/log_std Mean 0.988057 +trainer/policy/normal/log_std Std 0.337365 +trainer/policy/normal/log_std Max 1.81073 +trainer/policy/normal/log_std Min -1.27005 +eval/num steps total 528385 +eval/num paths total 532 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.181624 +eval/Actions Std 0.940817 +eval/Actions Max 0.999994 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64374 +time/logging (s) 0.00369971 +time/sampling batch (s) 0.281997 +time/saving (s) 0.00346035 +time/training (s) 6.99515 +time/epoch (s) 9.92805 +time/total (s) 5453.72 +Epoch -469 +---------------------------------- --------------- +2022-05-10 14:41:47.891991 PDT | [0] Epoch -468 finished +---------------------------------- --------------- +epoch -468 +replay_buffer/size 999033 +trainer/num train calls 533000 +trainer/Policy Loss -19.6553 +trainer/Log Pis Mean 23.755 +trainer/Log Pis Std 13.2507 +trainer/Log Pis Max 67.9988 +trainer/Log Pis Min -11.6442 +trainer/policy/mean Mean -0.0327713 +trainer/policy/mean Std 0.903735 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.80254 +trainer/policy/normal/std Std 0.626287 +trainer/policy/normal/std Max 5.47996 +trainer/policy/normal/std Min 0.379182 +trainer/policy/normal/log_std Mean 0.99597 +trainer/policy/normal/log_std Std 0.292707 +trainer/policy/normal/log_std Max 1.7011 +trainer/policy/normal/log_std Min -0.96974 +eval/num steps total 529385 +eval/num paths total 533 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.419626 +eval/Actions Std 0.732461 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59662 +time/logging (s) 0.00435029 +time/sampling batch (s) 0.534002 +time/saving (s) 0.00451888 +time/training (s) 8.37395 +time/epoch (s) 11.5134 +time/total (s) 5465.24 +Epoch -468 +---------------------------------- --------------- +2022-05-10 14:41:58.895604 PDT | [0] Epoch -467 finished +---------------------------------- --------------- +epoch -467 +replay_buffer/size 999033 +trainer/num train calls 534000 +trainer/Policy Loss -18.767 +trainer/Log Pis Mean 24.4313 +trainer/Log Pis Std 13.0463 +trainer/Log Pis Max 65.9118 +trainer/Log Pis Min -6.08599 +trainer/policy/mean Mean -0.0407051 +trainer/policy/mean Std 0.904609 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80513 +trainer/policy/normal/std Std 0.640756 +trainer/policy/normal/std Max 5.44383 +trainer/policy/normal/std Min 0.328678 +trainer/policy/normal/log_std Mean 0.993584 +trainer/policy/normal/log_std Std 0.310963 +trainer/policy/normal/log_std Max 1.69448 +trainer/policy/normal/log_std Min -1.11268 +eval/num steps total 530385 +eval/num paths total 534 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.045577 +eval/Actions Std 0.91247 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62159 +time/logging (s) 0.00390055 +time/sampling batch (s) 0.533234 +time/saving (s) 0.00369799 +time/training (s) 7.81767 +time/epoch (s) 10.9801 +time/total (s) 5476.22 +Epoch -467 +---------------------------------- --------------- +2022-05-10 14:42:08.865462 PDT | [0] Epoch -466 finished +---------------------------------- --------------- +epoch -466 +replay_buffer/size 999033 +trainer/num train calls 535000 +trainer/Policy Loss -18.9382 +trainer/Log Pis Mean 24.56 +trainer/Log Pis Std 12.5123 +trainer/Log Pis Max 59.2845 +trainer/Log Pis Min -8.87614 +trainer/policy/mean Mean -0.0471125 +trainer/policy/mean Std 0.905728 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79253 +trainer/policy/normal/std Std 0.654456 +trainer/policy/normal/std Max 4.91339 +trainer/policy/normal/std Min 0.271622 +trainer/policy/normal/log_std Mean 0.985444 +trainer/policy/normal/log_std Std 0.329708 +trainer/policy/normal/log_std Max 1.59197 +trainer/policy/normal/log_std Min -1.30334 +eval/num steps total 531385 +eval/num paths total 535 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.338478 +eval/Actions Std 0.850941 +eval/Actions Max 0.999995 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61226 +time/logging (s) 0.00378842 +time/sampling batch (s) 0.280765 +time/saving (s) 0.00348076 +time/training (s) 7.04755 +time/epoch (s) 9.94784 +time/total (s) 5486.17 +Epoch -466 +---------------------------------- --------------- +2022-05-10 14:42:18.599333 PDT | [0] Epoch -465 finished +---------------------------------- --------------- +epoch -465 +replay_buffer/size 999033 +trainer/num train calls 536000 +trainer/Policy Loss -19.2859 +trainer/Log Pis Mean 24.1222 +trainer/Log Pis Std 13.7229 +trainer/Log Pis Max 92.7611 +trainer/Log Pis Min -6.13213 +trainer/policy/mean Mean -0.06047 +trainer/policy/mean Std 0.900403 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77808 +trainer/policy/normal/std Std 0.645453 +trainer/policy/normal/std Max 5.49308 +trainer/policy/normal/std Min 0.316778 +trainer/policy/normal/log_std Mean 0.98181 +trainer/policy/normal/log_std Std 0.321346 +trainer/policy/normal/log_std Max 1.70349 +trainer/policy/normal/log_std Min -1.14955 +eval/num steps total 532385 +eval/num paths total 536 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.102096 +eval/Actions Std 0.915424 +eval/Actions Max 0.999976 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56486 +time/logging (s) 0.00384336 +time/sampling batch (s) 0.534964 +time/saving (s) 0.00349878 +time/training (s) 6.60478 +time/epoch (s) 9.71194 +time/total (s) 5495.88 +Epoch -465 +---------------------------------- --------------- +2022-05-10 14:42:28.154027 PDT | [0] Epoch -464 finished +---------------------------------- --------------- +epoch -464 +replay_buffer/size 999033 +trainer/num train calls 537000 +trainer/Policy Loss -20.1546 +trainer/Log Pis Mean 23.922 +trainer/Log Pis Std 12.8844 +trainer/Log Pis Max 61.9876 +trainer/Log Pis Min -8.45396 +trainer/policy/mean Mean -0.0801631 +trainer/policy/mean Std 0.904318 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78919 +trainer/policy/normal/std Std 0.631319 +trainer/policy/normal/std Max 5.6261 +trainer/policy/normal/std Min 0.289427 +trainer/policy/normal/log_std Mean 0.988186 +trainer/policy/normal/log_std Std 0.310942 +trainer/policy/normal/log_std Max 1.72742 +trainer/policy/normal/log_std Min -1.23985 +eval/num steps total 533385 +eval/num paths total 537 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.239291 +eval/Actions Std 0.961157 +eval/Actions Max 0.999962 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64414 +time/logging (s) 0.00376769 +time/sampling batch (s) 0.287531 +time/saving (s) 0.0034595 +time/training (s) 6.59245 +time/epoch (s) 9.53135 +time/total (s) 5505.42 +Epoch -464 +---------------------------------- --------------- +2022-05-10 14:42:37.553111 PDT | [0] Epoch -463 finished +---------------------------------- --------------- +epoch -463 +replay_buffer/size 999033 +trainer/num train calls 538000 +trainer/Policy Loss -19.9435 +trainer/Log Pis Mean 23.9308 +trainer/Log Pis Std 13.0131 +trainer/Log Pis Max 67.714 +trainer/Log Pis Min -5.85479 +trainer/policy/mean Mean -0.0395492 +trainer/policy/mean Std 0.902133 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80592 +trainer/policy/normal/std Std 0.640178 +trainer/policy/normal/std Max 5.34957 +trainer/policy/normal/std Min 0.340992 +trainer/policy/normal/log_std Mean 0.994509 +trainer/policy/normal/log_std Std 0.306964 +trainer/policy/normal/log_std Max 1.67702 +trainer/policy/normal/log_std Min -1.0759 +eval/num steps total 534385 +eval/num paths total 538 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.133684 +eval/Actions Std 0.913268 +eval/Actions Max 0.999992 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51401 +time/logging (s) 0.00382438 +time/sampling batch (s) 0.28663 +time/saving (s) 0.00349977 +time/training (s) 6.56906 +time/epoch (s) 9.37703 +time/total (s) 5514.8 +Epoch -463 +---------------------------------- --------------- +2022-05-10 14:42:47.952319 PDT | [0] Epoch -462 finished +---------------------------------- --------------- +epoch -462 +replay_buffer/size 999033 +trainer/num train calls 539000 +trainer/Policy Loss -19.6943 +trainer/Log Pis Mean 24.8274 +trainer/Log Pis Std 12.6248 +trainer/Log Pis Max 66.6448 +trainer/Log Pis Min -8.97844 +trainer/policy/mean Mean -0.00879921 +trainer/policy/mean Std 0.902895 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81708 +trainer/policy/normal/std Std 0.655662 +trainer/policy/normal/std Max 5.83393 +trainer/policy/normal/std Min 0.366805 +trainer/policy/normal/log_std Mean 0.996986 +trainer/policy/normal/log_std Std 0.313071 +trainer/policy/normal/log_std Max 1.76369 +trainer/policy/normal/log_std Min -1.00292 +eval/num steps total 535385 +eval/num paths total 539 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00679158 +eval/Actions Std 0.904762 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7041 +time/logging (s) 0.00399345 +time/sampling batch (s) 0.283239 +time/saving (s) 0.00415667 +time/training (s) 7.382 +time/epoch (s) 10.3775 +time/total (s) 5525.18 +Epoch -462 +---------------------------------- --------------- +2022-05-10 14:42:57.818781 PDT | [0] Epoch -461 finished +---------------------------------- --------------- +epoch -461 +replay_buffer/size 999033 +trainer/num train calls 540000 +trainer/Policy Loss -18.776 +trainer/Log Pis Mean 24.0826 +trainer/Log Pis Std 14.3096 +trainer/Log Pis Max 73.112 +trainer/Log Pis Min -7.91025 +trainer/policy/mean Mean -0.0343565 +trainer/policy/mean Std 0.905501 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86324 +trainer/policy/normal/std Std 0.673192 +trainer/policy/normal/std Max 5.55511 +trainer/policy/normal/std Min 0.281222 +trainer/policy/normal/log_std Mean 1.01113 +trainer/policy/normal/log_std Std 0.32567 +trainer/policy/normal/log_std Max 1.71472 +trainer/policy/normal/log_std Min -1.26861 +eval/num steps total 536385 +eval/num paths total 540 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0193623 +eval/Actions Std 0.912975 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50588 +time/logging (s) 0.00372817 +time/sampling batch (s) 0.280523 +time/saving (s) 0.00344007 +time/training (s) 7.05055 +time/epoch (s) 9.84412 +time/total (s) 5535.03 +Epoch -461 +---------------------------------- --------------- +2022-05-10 14:43:07.695764 PDT | [0] Epoch -460 finished +---------------------------------- --------------- +epoch -460 +replay_buffer/size 999033 +trainer/num train calls 541000 +trainer/Policy Loss -20.5209 +trainer/Log Pis Mean 24.885 +trainer/Log Pis Std 13.6351 +trainer/Log Pis Max 72.2959 +trainer/Log Pis Min -7.01844 +trainer/policy/mean Mean -0.0293486 +trainer/policy/mean Std 0.905006 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.87793 +trainer/policy/normal/std Std 0.660671 +trainer/policy/normal/std Max 6.4364 +trainer/policy/normal/std Min 0.322453 +trainer/policy/normal/log_std Mean 1.01891 +trainer/policy/normal/log_std Std 0.31177 +trainer/policy/normal/log_std Max 1.86197 +trainer/policy/normal/log_std Min -1.1318 +eval/num steps total 537385 +eval/num paths total 541 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.020679 +eval/Actions Std 0.898519 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75891 +time/logging (s) 0.00386952 +time/sampling batch (s) 0.280357 +time/saving (s) 0.00345251 +time/training (s) 6.80878 +time/epoch (s) 9.85537 +time/total (s) 5544.88 +Epoch -460 +---------------------------------- --------------- +2022-05-10 14:43:17.794365 PDT | [0] Epoch -459 finished +---------------------------------- --------------- +epoch -459 +replay_buffer/size 999033 +trainer/num train calls 542000 +trainer/Policy Loss -20.3101 +trainer/Log Pis Mean 24.3556 +trainer/Log Pis Std 12.633 +trainer/Log Pis Max 61.9306 +trainer/Log Pis Min -6.31779 +trainer/policy/mean Mean -0.0327054 +trainer/policy/mean Std 0.909932 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80605 +trainer/policy/normal/std Std 0.646543 +trainer/policy/normal/std Max 6.80454 +trainer/policy/normal/std Min 0.26206 +trainer/policy/normal/log_std Mean 0.993228 +trainer/policy/normal/log_std Std 0.313679 +trainer/policy/normal/log_std Max 1.91759 +trainer/policy/normal/log_std Min -1.33918 +eval/num steps total 538385 +eval/num paths total 542 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0619395 +eval/Actions Std 0.893424 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61835 +time/logging (s) 0.00372057 +time/sampling batch (s) 0.296235 +time/saving (s) 0.00348524 +time/training (s) 7.15399 +time/epoch (s) 10.0758 +time/total (s) 5554.96 +Epoch -459 +---------------------------------- --------------- +2022-05-10 14:43:28.463810 PDT | [0] Epoch -458 finished +---------------------------------- --------------- +epoch -458 +replay_buffer/size 999033 +trainer/num train calls 543000 +trainer/Policy Loss -20.2795 +trainer/Log Pis Mean 24.7964 +trainer/Log Pis Std 13.7216 +trainer/Log Pis Max 76.4337 +trainer/Log Pis Min -6.23737 +trainer/policy/mean Mean -0.0259536 +trainer/policy/mean Std 0.906653 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76875 +trainer/policy/normal/std Std 0.645361 +trainer/policy/normal/std Max 6.07816 +trainer/policy/normal/std Min 0.284248 +trainer/policy/normal/log_std Mean 0.976395 +trainer/policy/normal/log_std Std 0.333908 +trainer/policy/normal/log_std Max 1.8047 +trainer/policy/normal/log_std Min -1.25791 +eval/num steps total 539385 +eval/num paths total 543 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.449238 +eval/Actions Std 0.87724 +eval/Actions Max 0.999984 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79539 +time/logging (s) 0.00366444 +time/sampling batch (s) 0.286631 +time/saving (s) 0.00345726 +time/training (s) 7.55805 +time/epoch (s) 10.6472 +time/total (s) 5565.61 +Epoch -458 +---------------------------------- --------------- +2022-05-10 14:43:38.399696 PDT | [0] Epoch -457 finished +---------------------------------- --------------- +epoch -457 +replay_buffer/size 999033 +trainer/num train calls 544000 +trainer/Policy Loss -20.0178 +trainer/Log Pis Mean 24.3054 +trainer/Log Pis Std 12.8382 +trainer/Log Pis Max 61.8772 +trainer/Log Pis Min -14.5827 +trainer/policy/mean Mean -0.0375768 +trainer/policy/mean Std 0.907245 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80249 +trainer/policy/normal/std Std 0.684413 +trainer/policy/normal/std Max 7.10228 +trainer/policy/normal/std Min 0.312676 +trainer/policy/normal/log_std Mean 0.985978 +trainer/policy/normal/log_std Std 0.340249 +trainer/policy/normal/log_std Max 1.96042 +trainer/policy/normal/log_std Min -1.16259 +eval/num steps total 540385 +eval/num paths total 544 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0929035 +eval/Actions Std 0.90215 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55923 +time/logging (s) 0.00413322 +time/sampling batch (s) 0.281404 +time/saving (s) 0.00423571 +time/training (s) 7.06547 +time/epoch (s) 9.91448 +time/total (s) 5575.53 +Epoch -457 +---------------------------------- --------------- +2022-05-10 14:43:49.124702 PDT | [0] Epoch -456 finished +---------------------------------- --------------- +epoch -456 +replay_buffer/size 999033 +trainer/num train calls 545000 +trainer/Policy Loss -19.859 +trainer/Log Pis Mean 25.554 +trainer/Log Pis Std 13.1146 +trainer/Log Pis Max 69.8623 +trainer/Log Pis Min -2.38561 +trainer/policy/mean Mean -0.0320688 +trainer/policy/mean Std 0.903423 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83785 +trainer/policy/normal/std Std 0.679678 +trainer/policy/normal/std Max 5.78884 +trainer/policy/normal/std Min 0.248149 +trainer/policy/normal/log_std Mean 1.00031 +trainer/policy/normal/log_std Std 0.333239 +trainer/policy/normal/log_std Max 1.75593 +trainer/policy/normal/log_std Min -1.39373 +eval/num steps total 541385 +eval/num paths total 545 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.209566 +eval/Actions Std 0.73716 +eval/Actions Max 0.99999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68408 +time/logging (s) 0.00380228 +time/sampling batch (s) 0.281431 +time/saving (s) 0.00356867 +time/training (s) 7.72955 +time/epoch (s) 10.7024 +time/total (s) 5586.24 +Epoch -456 +---------------------------------- --------------- +2022-05-10 14:43:59.780222 PDT | [0] Epoch -455 finished +---------------------------------- --------------- +epoch -455 +replay_buffer/size 999033 +trainer/num train calls 546000 +trainer/Policy Loss -19.8238 +trainer/Log Pis Mean 23.7308 +trainer/Log Pis Std 12.796 +trainer/Log Pis Max 102.211 +trainer/Log Pis Min -4.9571 +trainer/policy/mean Mean -0.0236857 +trainer/policy/mean Std 0.903232 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77062 +trainer/policy/normal/std Std 0.660948 +trainer/policy/normal/std Max 5.49623 +trainer/policy/normal/std Min 0.237715 +trainer/policy/normal/log_std Mean 0.975857 +trainer/policy/normal/log_std Std 0.337326 +trainer/policy/normal/log_std Max 1.70406 +trainer/policy/normal/log_std Min -1.43668 +eval/num steps total 542385 +eval/num paths total 546 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.263141 +eval/Actions Std 0.822647 +eval/Actions Max 0.999991 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62202 +time/logging (s) 0.00386799 +time/sampling batch (s) 0.53118 +time/saving (s) 0.00342322 +time/training (s) 7.47323 +time/epoch (s) 10.6337 +time/total (s) 5596.87 +Epoch -455 +---------------------------------- --------------- +2022-05-10 14:44:10.983033 PDT | [0] Epoch -454 finished +---------------------------------- --------------- +epoch -454 +replay_buffer/size 999033 +trainer/num train calls 547000 +trainer/Policy Loss -19.541 +trainer/Log Pis Mean 23.6369 +trainer/Log Pis Std 12.6201 +trainer/Log Pis Max 63.1138 +trainer/Log Pis Min -10.4438 +trainer/policy/mean Mean -0.0465182 +trainer/policy/mean Std 0.908323 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.87199 +trainer/policy/normal/std Std 0.658972 +trainer/policy/normal/std Max 6.02143 +trainer/policy/normal/std Min 0.285024 +trainer/policy/normal/log_std Mean 1.01781 +trainer/policy/normal/log_std Std 0.305838 +trainer/policy/normal/log_std Max 1.79532 +trainer/policy/normal/log_std Min -1.25518 +eval/num steps total 543385 +eval/num paths total 547 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.306158 +eval/Actions Std 0.877223 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62856 +time/logging (s) 0.00374374 +time/sampling batch (s) 0.282445 +time/saving (s) 0.00339702 +time/training (s) 8.26277 +time/epoch (s) 11.1809 +time/total (s) 5608.06 +Epoch -454 +---------------------------------- --------------- +2022-05-10 14:44:21.209205 PDT | [0] Epoch -453 finished +---------------------------------- -------------- +epoch -453 +replay_buffer/size 999033 +trainer/num train calls 548000 +trainer/Policy Loss -18.9247 +trainer/Log Pis Mean 24.2481 +trainer/Log Pis Std 13.3863 +trainer/Log Pis Max 71.4499 +trainer/Log Pis Min -10.7982 +trainer/policy/mean Mean -0.0254703 +trainer/policy/mean Std 0.907041 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.78091 +trainer/policy/normal/std Std 0.646702 +trainer/policy/normal/std Max 5.8371 +trainer/policy/normal/std Min 0.315625 +trainer/policy/normal/log_std Mean 0.984154 +trainer/policy/normal/log_std Std 0.313123 +trainer/policy/normal/log_std Max 1.76423 +trainer/policy/normal/log_std Min -1.1532 +eval/num steps total 544385 +eval/num paths total 548 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.152417 +eval/Actions Std 0.932844 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60436 +time/logging (s) 0.003692 +time/sampling batch (s) 0.283081 +time/saving (s) 0.0034166 +time/training (s) 7.30984 +time/epoch (s) 10.2044 +time/total (s) 5618.26 +Epoch -453 +---------------------------------- -------------- +2022-05-10 14:44:31.837705 PDT | [0] Epoch -452 finished +---------------------------------- --------------- +epoch -452 +replay_buffer/size 999033 +trainer/num train calls 549000 +trainer/Policy Loss -19.9793 +trainer/Log Pis Mean 24.7482 +trainer/Log Pis Std 12.7966 +trainer/Log Pis Max 63.3488 +trainer/Log Pis Min -13.0921 +trainer/policy/mean Mean -0.0335959 +trainer/policy/mean Std 0.908145 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80734 +trainer/policy/normal/std Std 0.655761 +trainer/policy/normal/std Max 4.92912 +trainer/policy/normal/std Min 0.289734 +trainer/policy/normal/log_std Mean 0.99037 +trainer/policy/normal/log_std Std 0.332915 +trainer/policy/normal/log_std Max 1.59516 +trainer/policy/normal/log_std Min -1.23879 +eval/num steps total 545385 +eval/num paths total 549 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114494 +eval/Actions Std 0.915216 +eval/Actions Max 0.999995 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62511 +time/logging (s) 0.00410556 +time/sampling batch (s) 0.281541 +time/saving (s) 0.00406466 +time/training (s) 7.69212 +time/epoch (s) 10.607 +time/total (s) 5628.87 +Epoch -452 +---------------------------------- --------------- +2022-05-10 14:44:41.668527 PDT | [0] Epoch -451 finished +---------------------------------- --------------- +epoch -451 +replay_buffer/size 999033 +trainer/num train calls 550000 +trainer/Policy Loss -19.7996 +trainer/Log Pis Mean 25.8105 +trainer/Log Pis Std 12.8718 +trainer/Log Pis Max 65.9458 +trainer/Log Pis Min -5.04592 +trainer/policy/mean Mean -0.0438244 +trainer/policy/mean Std 0.903153 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82458 +trainer/policy/normal/std Std 0.648723 +trainer/policy/normal/std Max 5.26022 +trainer/policy/normal/std Min 0.339081 +trainer/policy/normal/log_std Mean 1.00085 +trainer/policy/normal/log_std Std 0.306919 +trainer/policy/normal/log_std Max 1.66017 +trainer/policy/normal/log_std Min -1.08152 +eval/num steps total 546385 +eval/num paths total 550 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0838934 +eval/Actions Std 0.899297 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54599 +time/logging (s) 0.00381668 +time/sampling batch (s) 0.286249 +time/saving (s) 0.00369034 +time/training (s) 6.96825 +time/epoch (s) 9.808 +time/total (s) 5638.68 +Epoch -451 +---------------------------------- --------------- +2022-05-10 14:44:51.017415 PDT | [0] Epoch -450 finished +---------------------------------- --------------- +epoch -450 +replay_buffer/size 999033 +trainer/num train calls 551000 +trainer/Policy Loss -19.6285 +trainer/Log Pis Mean 24.7494 +trainer/Log Pis Std 13.3707 +trainer/Log Pis Max 72.876 +trainer/Log Pis Min -5.85467 +trainer/policy/mean Mean -0.0503143 +trainer/policy/mean Std 0.907363 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.8751 +trainer/policy/normal/std Std 0.647665 +trainer/policy/normal/std Max 6.45239 +trainer/policy/normal/std Min 0.285326 +trainer/policy/normal/log_std Mean 1.01959 +trainer/policy/normal/log_std Std 0.305121 +trainer/policy/normal/log_std Max 1.86445 +trainer/policy/normal/log_std Min -1.25412 +eval/num steps total 547385 +eval/num paths total 551 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0216388 +eval/Actions Std 0.900405 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49456 +time/logging (s) 0.00376409 +time/sampling batch (s) 0.284418 +time/saving (s) 0.00350448 +time/training (s) 6.54044 +time/epoch (s) 9.32669 +time/total (s) 5648.01 +Epoch -450 +---------------------------------- --------------- +2022-05-10 14:45:01.831943 PDT | [0] Epoch -449 finished +---------------------------------- --------------- +epoch -449 +replay_buffer/size 999033 +trainer/num train calls 552000 +trainer/Policy Loss -18.7471 +trainer/Log Pis Mean 24.0467 +trainer/Log Pis Std 13.047 +trainer/Log Pis Max 60.0986 +trainer/Log Pis Min -5.85347 +trainer/policy/mean Mean -0.0223654 +trainer/policy/mean Std 0.902866 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81397 +trainer/policy/normal/std Std 0.618446 +trainer/policy/normal/std Max 5.58202 +trainer/policy/normal/std Min 0.310313 +trainer/policy/normal/log_std Mean 1.00138 +trainer/policy/normal/log_std Std 0.286612 +trainer/policy/normal/log_std Max 1.71955 +trainer/policy/normal/log_std Min -1.17017 +eval/num steps total 548385 +eval/num paths total 552 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00924167 +eval/Actions Std 0.954086 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6682 +time/logging (s) 0.00377396 +time/sampling batch (s) 0.541613 +time/saving (s) 0.00345157 +time/training (s) 7.57518 +time/epoch (s) 10.7922 +time/total (s) 5658.81 +Epoch -449 +---------------------------------- --------------- +2022-05-10 14:45:12.254588 PDT | [0] Epoch -448 finished +---------------------------------- --------------- +epoch -448 +replay_buffer/size 999033 +trainer/num train calls 553000 +trainer/Policy Loss -19.6932 +trainer/Log Pis Mean 25.1664 +trainer/Log Pis Std 13.558 +trainer/Log Pis Max 81.8622 +trainer/Log Pis Min -4.61402 +trainer/policy/mean Mean -0.0426342 +trainer/policy/mean Std 0.911298 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.80532 +trainer/policy/normal/std Std 0.652291 +trainer/policy/normal/std Max 6.05651 +trainer/policy/normal/std Min 0.247093 +trainer/policy/normal/log_std Mean 0.991521 +trainer/policy/normal/log_std Std 0.322508 +trainer/policy/normal/log_std Max 1.80113 +trainer/policy/normal/log_std Min -1.39799 +eval/num steps total 549385 +eval/num paths total 553 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0836549 +eval/Actions Std 0.904824 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46658 +time/logging (s) 0.00366514 +time/sampling batch (s) 0.784699 +time/saving (s) 0.00343565 +time/training (s) 7.14221 +time/epoch (s) 10.4006 +time/total (s) 5669.21 +Epoch -448 +---------------------------------- --------------- +2022-05-10 14:45:23.429130 PDT | [0] Epoch -447 finished +---------------------------------- --------------- +epoch -447 +replay_buffer/size 999033 +trainer/num train calls 554000 +trainer/Policy Loss -19.5923 +trainer/Log Pis Mean 24.3942 +trainer/Log Pis Std 12.7554 +trainer/Log Pis Max 68.9779 +trainer/Log Pis Min -9.57455 +trainer/policy/mean Mean -0.0288557 +trainer/policy/mean Std 0.908876 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79584 +trainer/policy/normal/std Std 0.64132 +trainer/policy/normal/std Max 5.96123 +trainer/policy/normal/std Min 0.259761 +trainer/policy/normal/log_std Mean 0.990674 +trainer/policy/normal/log_std Std 0.307585 +trainer/policy/normal/log_std Max 1.78528 +trainer/policy/normal/log_std Min -1.34799 +eval/num steps total 550385 +eval/num paths total 554 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11134 +eval/Actions Std 0.901 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61329 +time/logging (s) 0.00413953 +time/sampling batch (s) 0.283008 +time/saving (s) 0.00415739 +time/training (s) 8.24844 +time/epoch (s) 11.153 +time/total (s) 5680.37 +Epoch -447 +---------------------------------- --------------- +2022-05-10 14:45:33.386581 PDT | [0] Epoch -446 finished +---------------------------------- --------------- +epoch -446 +replay_buffer/size 999033 +trainer/num train calls 555000 +trainer/Policy Loss -18.8544 +trainer/Log Pis Mean 24.3121 +trainer/Log Pis Std 13.171 +trainer/Log Pis Max 68.8384 +trainer/Log Pis Min -9.30039 +trainer/policy/mean Mean -0.0207574 +trainer/policy/mean Std 0.90597 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85341 +trainer/policy/normal/std Std 0.673589 +trainer/policy/normal/std Max 6.9707 +trainer/policy/normal/std Min 0.325356 +trainer/policy/normal/log_std Mean 1.00889 +trainer/policy/normal/log_std Std 0.317101 +trainer/policy/normal/log_std Max 1.94172 +trainer/policy/normal/log_std Min -1.12284 +eval/num steps total 551385 +eval/num paths total 555 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.20514 +eval/Actions Std 0.898646 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5652 +time/logging (s) 0.00381547 +time/sampling batch (s) 0.281662 +time/saving (s) 0.00370528 +time/training (s) 7.08027 +time/epoch (s) 9.93466 +time/total (s) 5690.31 +Epoch -446 +---------------------------------- --------------- +2022-05-10 14:45:43.145493 PDT | [0] Epoch -445 finished +---------------------------------- --------------- +epoch -445 +replay_buffer/size 999033 +trainer/num train calls 556000 +trainer/Policy Loss -19.977 +trainer/Log Pis Mean 24.8064 +trainer/Log Pis Std 12.7508 +trainer/Log Pis Max 70.3604 +trainer/Log Pis Min -7.7329 +trainer/policy/mean Mean -0.00687991 +trainer/policy/mean Std 0.906294 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8316 +trainer/policy/normal/std Std 0.655368 +trainer/policy/normal/std Max 5.71989 +trainer/policy/normal/std Min 0.309636 +trainer/policy/normal/log_std Mean 1.00322 +trainer/policy/normal/log_std Std 0.308247 +trainer/policy/normal/log_std Max 1.74395 +trainer/policy/normal/log_std Min -1.17236 +eval/num steps total 552385 +eval/num paths total 556 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0486407 +eval/Actions Std 0.865283 +eval/Actions Max 0.999995 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62073 +time/logging (s) 0.00436998 +time/sampling batch (s) 0.285738 +time/saving (s) 0.00342931 +time/training (s) 6.82287 +time/epoch (s) 9.73713 +time/total (s) 5700.05 +Epoch -445 +---------------------------------- --------------- +2022-05-10 14:45:54.160800 PDT | [0] Epoch -444 finished +---------------------------------- --------------- +epoch -444 +replay_buffer/size 999033 +trainer/num train calls 557000 +trainer/Policy Loss -19.9067 +trainer/Log Pis Mean 24.1228 +trainer/Log Pis Std 13.0661 +trainer/Log Pis Max 73.3632 +trainer/Log Pis Min -5.46443 +trainer/policy/mean Mean -0.027743 +trainer/policy/mean Std 0.9076 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81072 +trainer/policy/normal/std Std 0.669547 +trainer/policy/normal/std Max 6.0465 +trainer/policy/normal/std Min 0.28174 +trainer/policy/normal/log_std Mean 0.991747 +trainer/policy/normal/log_std Std 0.326868 +trainer/policy/normal/log_std Max 1.79948 +trainer/policy/normal/log_std Min -1.26677 +eval/num steps total 553385 +eval/num paths total 557 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116062 +eval/Actions Std 0.917218 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73846 +time/logging (s) 0.00368041 +time/sampling batch (s) 0.539789 +time/saving (s) 0.00345072 +time/training (s) 7.70617 +time/epoch (s) 10.9915 +time/total (s) 5711.04 +Epoch -444 +---------------------------------- --------------- +2022-05-10 14:46:04.436633 PDT | [0] Epoch -443 finished +---------------------------------- --------------- +epoch -443 +replay_buffer/size 999033 +trainer/num train calls 558000 +trainer/Policy Loss -20.8276 +trainer/Log Pis Mean 25.1042 +trainer/Log Pis Std 13.2146 +trainer/Log Pis Max 73.7845 +trainer/Log Pis Min -8.58535 +trainer/policy/mean Mean -0.0614547 +trainer/policy/mean Std 0.912369 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83266 +trainer/policy/normal/std Std 0.634379 +trainer/policy/normal/std Max 6.95705 +trainer/policy/normal/std Min 0.408301 +trainer/policy/normal/log_std Mean 1.00614 +trainer/policy/normal/log_std Std 0.296489 +trainer/policy/normal/log_std Max 1.93976 +trainer/policy/normal/log_std Min -0.895752 +eval/num steps total 554385 +eval/num paths total 558 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.094579 +eval/Actions Std 0.918535 +eval/Actions Max 0.999995 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64546 +time/logging (s) 0.00371962 +time/sampling batch (s) 0.283944 +time/saving (s) 0.00356327 +time/training (s) 7.3171 +time/epoch (s) 10.2538 +time/total (s) 5721.3 +Epoch -443 +---------------------------------- --------------- +2022-05-10 14:46:13.797819 PDT | [0] Epoch -442 finished +---------------------------------- --------------- +epoch -442 +replay_buffer/size 999033 +trainer/num train calls 559000 +trainer/Policy Loss -20.2375 +trainer/Log Pis Mean 24.532 +trainer/Log Pis Std 13.5119 +trainer/Log Pis Max 74.5819 +trainer/Log Pis Min -5.89967 +trainer/policy/mean Mean -0.0343902 +trainer/policy/mean Std 0.906072 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.76513 +trainer/policy/normal/std Std 0.658857 +trainer/policy/normal/std Max 5.57114 +trainer/policy/normal/std Min 0.289756 +trainer/policy/normal/log_std Mean 0.973788 +trainer/policy/normal/log_std Std 0.337655 +trainer/policy/normal/log_std Max 1.7176 +trainer/policy/normal/log_std Min -1.23872 +eval/num steps total 555385 +eval/num paths total 559 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.204207 +eval/Actions Std 0.958227 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51233 +time/logging (s) 0.00374238 +time/sampling batch (s) 0.284164 +time/saving (s) 0.00345322 +time/training (s) 6.5354 +time/epoch (s) 9.3391 +time/total (s) 5730.64 +Epoch -442 +---------------------------------- --------------- +2022-05-10 14:46:24.034876 PDT | [0] Epoch -441 finished +---------------------------------- --------------- +epoch -441 +replay_buffer/size 999033 +trainer/num train calls 560000 +trainer/Policy Loss -20.0138 +trainer/Log Pis Mean 23.3323 +trainer/Log Pis Std 13.1474 +trainer/Log Pis Max 65.5441 +trainer/Log Pis Min -11.7995 +trainer/policy/mean Mean -0.031684 +trainer/policy/mean Std 0.900053 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74484 +trainer/policy/normal/std Std 0.663417 +trainer/policy/normal/std Max 5.58513 +trainer/policy/normal/std Min 0.327184 +trainer/policy/normal/log_std Mean 0.967021 +trainer/policy/normal/log_std Std 0.3305 +trainer/policy/normal/log_std Max 1.72011 +trainer/policy/normal/log_std Min -1.11723 +eval/num steps total 556385 +eval/num paths total 560 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.179608 +eval/Actions Std 0.897273 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51105 +time/logging (s) 0.00393961 +time/sampling batch (s) 0.282825 +time/saving (s) 0.00415654 +time/training (s) 7.41326 +time/epoch (s) 10.2152 +time/total (s) 5740.86 +Epoch -441 +---------------------------------- --------------- +2022-05-10 14:46:34.468117 PDT | [0] Epoch -440 finished +---------------------------------- --------------- +epoch -440 +replay_buffer/size 999033 +trainer/num train calls 561000 +trainer/Policy Loss -19.758 +trainer/Log Pis Mean 24.1425 +trainer/Log Pis Std 13.4543 +trainer/Log Pis Max 80.7065 +trainer/Log Pis Min -11.0108 +trainer/policy/mean Mean -0.0132038 +trainer/policy/mean Std 0.904775 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79959 +trainer/policy/normal/std Std 0.646308 +trainer/policy/normal/std Max 5.21574 +trainer/policy/normal/std Min 0.320769 +trainer/policy/normal/log_std Mean 0.989691 +trainer/policy/normal/log_std Std 0.321534 +trainer/policy/normal/log_std Max 1.65168 +trainer/policy/normal/log_std Min -1.13703 +eval/num steps total 557385 +eval/num paths total 561 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.019618 +eval/Actions Std 0.914546 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48875 +time/logging (s) 0.00375755 +time/sampling batch (s) 0.282185 +time/saving (s) 0.00346478 +time/training (s) 7.63233 +time/epoch (s) 10.4105 +time/total (s) 5751.27 +Epoch -440 +---------------------------------- --------------- +2022-05-10 14:46:44.622238 PDT | [0] Epoch -439 finished +---------------------------------- --------------- +epoch -439 +replay_buffer/size 999033 +trainer/num train calls 562000 +trainer/Policy Loss -19.4536 +trainer/Log Pis Mean 24.3285 +trainer/Log Pis Std 12.7335 +trainer/Log Pis Max 63.2133 +trainer/Log Pis Min -9.8616 +trainer/policy/mean Mean -0.032037 +trainer/policy/mean Std 0.90287 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8081 +trainer/policy/normal/std Std 0.654307 +trainer/policy/normal/std Max 6.99257 +trainer/policy/normal/std Min 0.25109 +trainer/policy/normal/log_std Mean 0.992412 +trainer/policy/normal/log_std Std 0.323745 +trainer/policy/normal/log_std Max 1.94485 +trainer/policy/normal/log_std Min -1.38194 +eval/num steps total 558385 +eval/num paths total 562 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.131001 +eval/Actions Std 0.889665 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66374 +time/logging (s) 0.00379409 +time/sampling batch (s) 0.282715 +time/saving (s) 0.00348247 +time/training (s) 7.17823 +time/epoch (s) 10.132 +time/total (s) 5761.41 +Epoch -439 +---------------------------------- --------------- +2022-05-10 14:46:54.819528 PDT | [0] Epoch -438 finished +---------------------------------- --------------- +epoch -438 +replay_buffer/size 999033 +trainer/num train calls 563000 +trainer/Policy Loss -19.6916 +trainer/Log Pis Mean 24.8021 +trainer/Log Pis Std 13.4009 +trainer/Log Pis Max 75.1343 +trainer/Log Pis Min -12.3012 +trainer/policy/mean Mean -0.0333084 +trainer/policy/mean Std 0.906138 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80324 +trainer/policy/normal/std Std 0.670783 +trainer/policy/normal/std Max 7.00637 +trainer/policy/normal/std Min 0.344577 +trainer/policy/normal/log_std Mean 0.987834 +trainer/policy/normal/log_std Std 0.334053 +trainer/policy/normal/log_std Max 1.94682 +trainer/policy/normal/log_std Min -1.06544 +eval/num steps total 559385 +eval/num paths total 563 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.313333 +eval/Actions Std 0.84499 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67577 +time/logging (s) 0.00375077 +time/sampling batch (s) 0.281964 +time/saving (s) 0.00340011 +time/training (s) 7.21033 +time/epoch (s) 10.1752 +time/total (s) 5771.58 +Epoch -438 +---------------------------------- --------------- +2022-05-10 14:47:05.245947 PDT | [0] Epoch -437 finished +---------------------------------- --------------- +epoch -437 +replay_buffer/size 999033 +trainer/num train calls 564000 +trainer/Policy Loss -20.0393 +trainer/Log Pis Mean 24.422 +trainer/Log Pis Std 13.2157 +trainer/Log Pis Max 69.6477 +trainer/Log Pis Min -13.0021 +trainer/policy/mean Mean -0.0268668 +trainer/policy/mean Std 0.907518 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.86048 +trainer/policy/normal/std Std 0.672787 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.330581 +trainer/policy/normal/log_std Mean 1.00995 +trainer/policy/normal/log_std Std 0.327448 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.1069 +eval/num steps total 560385 +eval/num paths total 564 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.185018 +eval/Actions Std 0.908878 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.35285 +time/logging (s) 0.00372895 +time/sampling batch (s) 0.278872 +time/saving (s) 0.00358771 +time/training (s) 7.76573 +time/epoch (s) 10.4048 +time/total (s) 5781.99 +Epoch -437 +---------------------------------- --------------- +2022-05-10 14:47:15.511878 PDT | [0] Epoch -436 finished +---------------------------------- --------------- +epoch -436 +replay_buffer/size 999033 +trainer/num train calls 565000 +trainer/Policy Loss -20.1429 +trainer/Log Pis Mean 24.0229 +trainer/Log Pis Std 12.9352 +trainer/Log Pis Max 79.3588 +trainer/Log Pis Min -6.54349 +trainer/policy/mean Mean -0.0478413 +trainer/policy/mean Std 0.900845 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81473 +trainer/policy/normal/std Std 0.69563 +trainer/policy/normal/std Max 6.71215 +trainer/policy/normal/std Min 0.358331 +trainer/policy/normal/log_std Mean 0.989622 +trainer/policy/normal/log_std Std 0.342092 +trainer/policy/normal/log_std Max 1.90392 +trainer/policy/normal/log_std Min -1.0263 +eval/num steps total 561385 +eval/num paths total 565 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.218874 +eval/Actions Std 0.827233 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58213 +time/logging (s) 0.00419015 +time/sampling batch (s) 0.781313 +time/saving (s) 0.00419559 +time/training (s) 6.87251 +time/epoch (s) 10.2443 +time/total (s) 5792.24 +Epoch -436 +---------------------------------- --------------- +2022-05-10 14:47:25.285050 PDT | [0] Epoch -435 finished +---------------------------------- --------------- +epoch -435 +replay_buffer/size 999033 +trainer/num train calls 566000 +trainer/Policy Loss -20.863 +trainer/Log Pis Mean 25.2878 +trainer/Log Pis Std 12.6475 +trainer/Log Pis Max 69.2266 +trainer/Log Pis Min -3.81685 +trainer/policy/mean Mean -0.0182383 +trainer/policy/mean Std 0.907523 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81341 +trainer/policy/normal/std Std 0.658891 +trainer/policy/normal/std Max 6.99505 +trainer/policy/normal/std Min 0.294057 +trainer/policy/normal/log_std Mean 0.994786 +trainer/policy/normal/log_std Std 0.318207 +trainer/policy/normal/log_std Max 1.9452 +trainer/policy/normal/log_std Min -1.22398 +eval/num steps total 562385 +eval/num paths total 566 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.127494 +eval/Actions Std 0.905767 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49259 +time/logging (s) 0.00366919 +time/sampling batch (s) 0.283299 +time/saving (s) 0.00359327 +time/training (s) 6.96681 +time/epoch (s) 9.74996 +time/total (s) 5801.99 +Epoch -435 +---------------------------------- --------------- +2022-05-10 14:47:34.970715 PDT | [0] Epoch -434 finished +---------------------------------- --------------- +epoch -434 +replay_buffer/size 999033 +trainer/num train calls 567000 +trainer/Policy Loss -19.1846 +trainer/Log Pis Mean 23.7223 +trainer/Log Pis Std 13.0444 +trainer/Log Pis Max 70.6253 +trainer/Log Pis Min -13.497 +trainer/policy/mean Mean -0.0259224 +trainer/policy/mean Std 0.907886 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.82591 +trainer/policy/normal/std Std 0.652673 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.262223 +trainer/policy/normal/log_std Mean 0.999848 +trainer/policy/normal/log_std Std 0.317081 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.33856 +eval/num steps total 563385 +eval/num paths total 567 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.152036 +eval/Actions Std 0.925284 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6958 +time/logging (s) 0.00374921 +time/sampling batch (s) 0.283219 +time/saving (s) 0.00343007 +time/training (s) 6.67761 +time/epoch (s) 9.66381 +time/total (s) 5811.66 +Epoch -434 +---------------------------------- --------------- +2022-05-10 14:47:44.328288 PDT | [0] Epoch -433 finished +---------------------------------- --------------- +epoch -433 +replay_buffer/size 999033 +trainer/num train calls 568000 +trainer/Policy Loss -19.6406 +trainer/Log Pis Mean 24.7333 +trainer/Log Pis Std 13.0783 +trainer/Log Pis Max 72.1259 +trainer/Log Pis Min -3.08453 +trainer/policy/mean Mean -0.0466356 +trainer/policy/mean Std 0.904233 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.86317 +trainer/policy/normal/std Std 0.661175 +trainer/policy/normal/std Max 5.52469 +trainer/policy/normal/std Min 0.30987 +trainer/policy/normal/log_std Mean 1.01272 +trainer/policy/normal/log_std Std 0.317801 +trainer/policy/normal/log_std Max 1.70923 +trainer/policy/normal/log_std Min -1.1716 +eval/num steps total 564385 +eval/num paths total 568 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.131014 +eval/Actions Std 0.857139 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73888 +time/logging (s) 0.00379452 +time/sampling batch (s) 0.281892 +time/saving (s) 0.00345716 +time/training (s) 6.30765 +time/epoch (s) 9.33568 +time/total (s) 5821 +Epoch -433 +---------------------------------- --------------- +2022-05-10 14:47:54.975224 PDT | [0] Epoch -432 finished +---------------------------------- --------------- +epoch -432 +replay_buffer/size 999033 +trainer/num train calls 569000 +trainer/Policy Loss -19.8535 +trainer/Log Pis Mean 25.1928 +trainer/Log Pis Std 13.4733 +trainer/Log Pis Max 77.7832 +trainer/Log Pis Min -7.73288 +trainer/policy/mean Mean -0.0456823 +trainer/policy/mean Std 0.908272 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.83032 +trainer/policy/normal/std Std 0.653744 +trainer/policy/normal/std Max 5.75524 +trainer/policy/normal/std Min 0.343919 +trainer/policy/normal/log_std Mean 1.0024 +trainer/policy/normal/log_std Std 0.310266 +trainer/policy/normal/log_std Max 1.75011 +trainer/policy/normal/log_std Min -1.06735 +eval/num steps total 565385 +eval/num paths total 569 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167691 +eval/Actions Std 0.932381 +eval/Actions Max 0.999991 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39275 +time/logging (s) 0.00380942 +time/sampling batch (s) 0.781181 +time/saving (s) 0.00344218 +time/training (s) 7.44387 +time/epoch (s) 10.6251 +time/total (s) 5831.63 +Epoch -432 +---------------------------------- --------------- +2022-05-10 14:48:04.331763 PDT | [0] Epoch -431 finished +---------------------------------- --------------- +epoch -431 +replay_buffer/size 999033 +trainer/num train calls 570000 +trainer/Policy Loss -18.7309 +trainer/Log Pis Mean 24.7756 +trainer/Log Pis Std 13.2514 +trainer/Log Pis Max 69.6205 +trainer/Log Pis Min -3.6292 +trainer/policy/mean Mean -0.0287467 +trainer/policy/mean Std 0.907942 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.78827 +trainer/policy/normal/std Std 0.650838 +trainer/policy/normal/std Max 5.31595 +trainer/policy/normal/std Min 0.299182 +trainer/policy/normal/log_std Mean 0.983954 +trainer/policy/normal/log_std Std 0.330082 +trainer/policy/normal/log_std Max 1.67071 +trainer/policy/normal/log_std Min -1.2067 +eval/num steps total 566385 +eval/num paths total 570 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0750452 +eval/Actions Std 0.87886 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68441 +time/logging (s) 0.00418153 +time/sampling batch (s) 0.279286 +time/saving (s) 0.00405613 +time/training (s) 6.36323 +time/epoch (s) 9.33517 +time/total (s) 5840.96 +Epoch -431 +---------------------------------- --------------- +2022-05-10 14:48:14.665507 PDT | [0] Epoch -430 finished +---------------------------------- --------------- +epoch -430 +replay_buffer/size 999033 +trainer/num train calls 571000 +trainer/Policy Loss -18.8706 +trainer/Log Pis Mean 23.8232 +trainer/Log Pis Std 12.5311 +trainer/Log Pis Max 71.1736 +trainer/Log Pis Min -6.29639 +trainer/policy/mean Mean -0.0304259 +trainer/policy/mean Std 0.902197 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.74832 +trainer/policy/normal/std Std 0.632122 +trainer/policy/normal/std Max 5.33299 +trainer/policy/normal/std Min 0.292175 +trainer/policy/normal/log_std Mean 0.972937 +trainer/policy/normal/log_std Std 0.311194 +trainer/policy/normal/log_std Max 1.67391 +trainer/policy/normal/log_std Min -1.2304 +eval/num steps total 567385 +eval/num paths total 571 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.359106 +eval/Actions Std 0.864454 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.608 +time/logging (s) 0.00379059 +time/sampling batch (s) 0.779158 +time/saving (s) 0.00377976 +time/training (s) 6.91636 +time/epoch (s) 10.3111 +time/total (s) 5851.28 +Epoch -430 +---------------------------------- --------------- +2022-05-10 14:48:23.463375 PDT | [0] Epoch -429 finished +---------------------------------- --------------- +epoch -429 +replay_buffer/size 999033 +trainer/num train calls 572000 +trainer/Policy Loss -19.2107 +trainer/Log Pis Mean 24.1849 +trainer/Log Pis Std 13.1848 +trainer/Log Pis Max 63.5917 +trainer/Log Pis Min -4.64193 +trainer/policy/mean Mean -0.0477758 +trainer/policy/mean Std 0.903979 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78882 +trainer/policy/normal/std Std 0.662151 +trainer/policy/normal/std Max 5.25709 +trainer/policy/normal/std Min 0.30482 +trainer/policy/normal/log_std Mean 0.983237 +trainer/policy/normal/log_std Std 0.332443 +trainer/policy/normal/log_std Max 1.65958 +trainer/policy/normal/log_std Min -1.18803 +eval/num steps total 568385 +eval/num paths total 572 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0120239 +eval/Actions Std 0.907635 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72578 +time/logging (s) 0.00374923 +time/sampling batch (s) 0.27933 +time/saving (s) 0.00347506 +time/training (s) 5.76357 +time/epoch (s) 8.7759 +time/total (s) 5860.06 +Epoch -429 +---------------------------------- --------------- +2022-05-10 14:48:33.314878 PDT | [0] Epoch -428 finished +---------------------------------- --------------- +epoch -428 +replay_buffer/size 999033 +trainer/num train calls 573000 +trainer/Policy Loss -19.1487 +trainer/Log Pis Mean 25.5092 +trainer/Log Pis Std 14.4221 +trainer/Log Pis Max 83.7393 +trainer/Log Pis Min -13.3483 +trainer/policy/mean Mean -0.0412843 +trainer/policy/mean Std 0.903914 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80687 +trainer/policy/normal/std Std 0.675544 +trainer/policy/normal/std Max 7.30521 +trainer/policy/normal/std Min 0.243173 +trainer/policy/normal/log_std Mean 0.988634 +trainer/policy/normal/log_std Std 0.33757 +trainer/policy/normal/log_std Max 1.98859 +trainer/policy/normal/log_std Min -1.41398 +eval/num steps total 569385 +eval/num paths total 573 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0398856 +eval/Actions Std 0.914834 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40914 +time/logging (s) 0.0037404 +time/sampling batch (s) 0.279944 +time/saving (s) 0.00344306 +time/training (s) 7.13344 +time/epoch (s) 9.82971 +time/total (s) 5869.89 +Epoch -428 +---------------------------------- --------------- +2022-05-10 14:48:42.836710 PDT | [0] Epoch -427 finished +---------------------------------- --------------- +epoch -427 +replay_buffer/size 999033 +trainer/num train calls 574000 +trainer/Policy Loss -18.8744 +trainer/Log Pis Mean 24.0494 +trainer/Log Pis Std 13.2575 +trainer/Log Pis Max 72.6369 +trainer/Log Pis Min -8.18571 +trainer/policy/mean Mean -0.0354528 +trainer/policy/mean Std 0.90849 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.82204 +trainer/policy/normal/std Std 0.66091 +trainer/policy/normal/std Max 5.65671 +trainer/policy/normal/std Min 0.275274 +trainer/policy/normal/log_std Mean 0.995343 +trainer/policy/normal/log_std Std 0.333058 +trainer/policy/normal/log_std Max 1.73284 +trainer/policy/normal/log_std Min -1.28999 +eval/num steps total 570385 +eval/num paths total 574 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132521 +eval/Actions Std 0.873275 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64637 +time/logging (s) 0.00379904 +time/sampling batch (s) 0.279624 +time/saving (s) 0.00343641 +time/training (s) 6.56686 +time/epoch (s) 9.50009 +time/total (s) 5879.39 +Epoch -427 +---------------------------------- --------------- +2022-05-10 14:48:53.878315 PDT | [0] Epoch -426 finished +---------------------------------- --------------- +epoch -426 +replay_buffer/size 999033 +trainer/num train calls 575000 +trainer/Policy Loss -18.78 +trainer/Log Pis Mean 24.9156 +trainer/Log Pis Std 13.7206 +trainer/Log Pis Max 70.983 +trainer/Log Pis Min -8.42451 +trainer/policy/mean Mean -0.03423 +trainer/policy/mean Std 0.90973 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81485 +trainer/policy/normal/std Std 0.652652 +trainer/policy/normal/std Max 5.57636 +trainer/policy/normal/std Min 0.287967 +trainer/policy/normal/log_std Mean 0.995013 +trainer/policy/normal/log_std Std 0.321811 +trainer/policy/normal/log_std Max 1.71854 +trainer/policy/normal/log_std Min -1.24491 +eval/num steps total 571385 +eval/num paths total 575 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0106205 +eval/Actions Std 0.903229 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59666 +time/logging (s) 0.00372098 +time/sampling batch (s) 0.284324 +time/saving (s) 0.00346648 +time/training (s) 8.13136 +time/epoch (s) 11.0195 +time/total (s) 5890.41 +Epoch -426 +---------------------------------- --------------- +2022-05-10 14:49:05.703840 PDT | [0] Epoch -425 finished +---------------------------------- --------------- +epoch -425 +replay_buffer/size 999033 +trainer/num train calls 576000 +trainer/Policy Loss -19.2632 +trainer/Log Pis Mean 24.2924 +trainer/Log Pis Std 13.672 +trainer/Log Pis Max 80.5458 +trainer/Log Pis Min -8.15652 +trainer/policy/mean Mean -0.0352923 +trainer/policy/mean Std 0.906445 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7707 +trainer/policy/normal/std Std 0.65912 +trainer/policy/normal/std Max 6.08765 +trainer/policy/normal/std Min 0.312431 +trainer/policy/normal/log_std Mean 0.978158 +trainer/policy/normal/log_std Std 0.323316 +trainer/policy/normal/log_std Max 1.80626 +trainer/policy/normal/log_std Min -1.16337 +eval/num steps total 572385 +eval/num paths total 576 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.267741 +eval/Actions Std 0.764633 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59432 +time/logging (s) 0.00414399 +time/sampling batch (s) 0.780315 +time/saving (s) 0.00412756 +time/training (s) 8.4211 +time/epoch (s) 11.804 +time/total (s) 5902.22 +Epoch -425 +---------------------------------- --------------- +2022-05-10 14:49:15.270041 PDT | [0] Epoch -424 finished +---------------------------------- --------------- +epoch -424 +replay_buffer/size 999033 +trainer/num train calls 577000 +trainer/Policy Loss -19.472 +trainer/Log Pis Mean 24.5411 +trainer/Log Pis Std 13.5936 +trainer/Log Pis Max 66.9046 +trainer/Log Pis Min -5.65395 +trainer/policy/mean Mean -0.0212023 +trainer/policy/mean Std 0.903049 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81778 +trainer/policy/normal/std Std 0.670497 +trainer/policy/normal/std Max 6.63892 +trainer/policy/normal/std Min 0.332131 +trainer/policy/normal/log_std Mean 0.994785 +trainer/policy/normal/log_std Std 0.325025 +trainer/policy/normal/log_std Max 1.89295 +trainer/policy/normal/log_std Min -1.10223 +eval/num steps total 573385 +eval/num paths total 577 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.381638 +eval/Actions Std 0.75147 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44312 +time/logging (s) 0.00381372 +time/sampling batch (s) 0.280224 +time/saving (s) 0.00355677 +time/training (s) 6.81271 +time/epoch (s) 9.54342 +time/total (s) 5911.77 +Epoch -424 +---------------------------------- --------------- +2022-05-10 14:49:25.919636 PDT | [0] Epoch -423 finished +---------------------------------- --------------- +epoch -423 +replay_buffer/size 999033 +trainer/num train calls 578000 +trainer/Policy Loss -19.3404 +trainer/Log Pis Mean 23.4528 +trainer/Log Pis Std 13.688 +trainer/Log Pis Max 68.4596 +trainer/Log Pis Min -14.2064 +trainer/policy/mean Mean -0.0503766 +trainer/policy/mean Std 0.901288 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79369 +trainer/policy/normal/std Std 0.632778 +trainer/policy/normal/std Max 5.89243 +trainer/policy/normal/std Min 0.248653 +trainer/policy/normal/log_std Mean 0.99109 +trainer/policy/normal/log_std Std 0.302405 +trainer/policy/normal/log_std Max 1.77367 +trainer/policy/normal/log_std Min -1.3917 +eval/num steps total 574385 +eval/num paths total 578 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.196426 +eval/Actions Std 0.957964 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59654 +time/logging (s) 0.0036883 +time/sampling batch (s) 0.534226 +time/saving (s) 0.00345882 +time/training (s) 7.4893 +time/epoch (s) 10.6272 +time/total (s) 5922.4 +Epoch -423 +---------------------------------- --------------- +2022-05-10 14:49:35.757641 PDT | [0] Epoch -422 finished +---------------------------------- --------------- +epoch -422 +replay_buffer/size 999033 +trainer/num train calls 579000 +trainer/Policy Loss -19.5151 +trainer/Log Pis Mean 23.7554 +trainer/Log Pis Std 13.5228 +trainer/Log Pis Max 66.9038 +trainer/Log Pis Min -9.06831 +trainer/policy/mean Mean -0.044238 +trainer/policy/mean Std 0.904607 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7989 +trainer/policy/normal/std Std 0.643482 +trainer/policy/normal/std Max 5.22561 +trainer/policy/normal/std Min 0.262071 +trainer/policy/normal/log_std Mean 0.989297 +trainer/policy/normal/log_std Std 0.32244 +trainer/policy/normal/log_std Max 1.65357 +trainer/policy/normal/log_std Min -1.33914 +eval/num steps total 575385 +eval/num paths total 579 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00358463 +eval/Actions Std 0.984712 +eval/Actions Max 0.999987 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60734 +time/logging (s) 0.00381976 +time/sampling batch (s) 0.282273 +time/saving (s) 0.00355659 +time/training (s) 6.91909 +time/epoch (s) 9.81609 +time/total (s) 5932.22 +Epoch -422 +---------------------------------- --------------- +2022-05-10 14:49:45.553354 PDT | [0] Epoch -421 finished +---------------------------------- --------------- +epoch -421 +replay_buffer/size 999033 +trainer/num train calls 580000 +trainer/Policy Loss -20.2097 +trainer/Log Pis Mean 25.3734 +trainer/Log Pis Std 13.1885 +trainer/Log Pis Max 64.2569 +trainer/Log Pis Min -7.60159 +trainer/policy/mean Mean -0.0294419 +trainer/policy/mean Std 0.913872 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82468 +trainer/policy/normal/std Std 0.646989 +trainer/policy/normal/std Max 6.12694 +trainer/policy/normal/std Min 0.339824 +trainer/policy/normal/log_std Mean 1.00113 +trainer/policy/normal/log_std Std 0.307172 +trainer/policy/normal/log_std Max 1.8127 +trainer/policy/normal/log_std Min -1.07933 +eval/num steps total 576385 +eval/num paths total 580 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114076 +eval/Actions Std 0.890473 +eval/Actions Max 0.999991 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58582 +time/logging (s) 0.00368889 +time/sampling batch (s) 0.539369 +time/saving (s) 0.00345533 +time/training (s) 6.6408 +time/epoch (s) 9.77313 +time/total (s) 5941.99 +Epoch -421 +---------------------------------- --------------- +2022-05-10 14:49:55.611134 PDT | [0] Epoch -420 finished +---------------------------------- --------------- +epoch -420 +replay_buffer/size 999033 +trainer/num train calls 581000 +trainer/Policy Loss -19.5744 +trainer/Log Pis Mean 24.8462 +trainer/Log Pis Std 12.9771 +trainer/Log Pis Max 66.8933 +trainer/Log Pis Min -7.98736 +trainer/policy/mean Mean -0.042574 +trainer/policy/mean Std 0.907483 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77785 +trainer/policy/normal/std Std 0.648032 +trainer/policy/normal/std Max 5.72568 +trainer/policy/normal/std Min 0.29489 +trainer/policy/normal/log_std Mean 0.981913 +trainer/policy/normal/log_std Std 0.320072 +trainer/policy/normal/log_std Max 1.74496 +trainer/policy/normal/log_std Min -1.22115 +eval/num steps total 577385 +eval/num paths total 581 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00458105 +eval/Actions Std 0.901558 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80061 +time/logging (s) 0.00402791 +time/sampling batch (s) 0.534625 +time/saving (s) 0.00375413 +time/training (s) 6.693 +time/epoch (s) 10.036 +time/total (s) 5952.03 +Epoch -420 +---------------------------------- --------------- +2022-05-10 14:50:06.446397 PDT | [0] Epoch -419 finished +---------------------------------- --------------- +epoch -419 +replay_buffer/size 999033 +trainer/num train calls 582000 +trainer/Policy Loss -19.2371 +trainer/Log Pis Mean 25.9258 +trainer/Log Pis Std 14.4747 +trainer/Log Pis Max 77.1571 +trainer/Log Pis Min -5.75325 +trainer/policy/mean Mean -0.0245418 +trainer/policy/mean Std 0.907399 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.85937 +trainer/policy/normal/std Std 0.657328 +trainer/policy/normal/std Max 6.09055 +trainer/policy/normal/std Min 0.241081 +trainer/policy/normal/log_std Mean 1.01252 +trainer/policy/normal/log_std Std 0.311957 +trainer/policy/normal/log_std Max 1.80674 +trainer/policy/normal/log_std Min -1.42262 +eval/num steps total 578385 +eval/num paths total 582 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0866916 +eval/Actions Std 0.897785 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7372 +time/logging (s) 0.00436638 +time/sampling batch (s) 0.787018 +time/saving (s) 0.00438036 +time/training (s) 7.28005 +time/epoch (s) 10.813 +time/total (s) 5962.85 +Epoch -419 +---------------------------------- --------------- +2022-05-10 14:50:16.296731 PDT | [0] Epoch -418 finished +---------------------------------- --------------- +epoch -418 +replay_buffer/size 999033 +trainer/num train calls 583000 +trainer/Policy Loss -19.7028 +trainer/Log Pis Mean 25.8486 +trainer/Log Pis Std 13.0227 +trainer/Log Pis Max 65.7749 +trainer/Log Pis Min -13.2732 +trainer/policy/mean Mean -0.0599651 +trainer/policy/mean Std 0.907521 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83302 +trainer/policy/normal/std Std 0.649702 +trainer/policy/normal/std Max 6.42953 +trainer/policy/normal/std Min 0.273139 +trainer/policy/normal/log_std Mean 1.00296 +trainer/policy/normal/log_std Std 0.314817 +trainer/policy/normal/log_std Max 1.8609 +trainer/policy/normal/log_std Min -1.29777 +eval/num steps total 579385 +eval/num paths total 583 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119131 +eval/Actions Std 0.909689 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59437 +time/logging (s) 0.00431231 +time/sampling batch (s) 0.27973 +time/saving (s) 0.00449782 +time/training (s) 6.94474 +time/epoch (s) 9.82764 +time/total (s) 5972.68 +Epoch -418 +---------------------------------- --------------- +2022-05-10 14:50:25.495089 PDT | [0] Epoch -417 finished +---------------------------------- --------------- +epoch -417 +replay_buffer/size 999033 +trainer/num train calls 584000 +trainer/Policy Loss -20.1896 +trainer/Log Pis Mean 24.9865 +trainer/Log Pis Std 13.0927 +trainer/Log Pis Max 71.4701 +trainer/Log Pis Min -7.71654 +trainer/policy/mean Mean 0.00312999 +trainer/policy/mean Std 0.90706 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.86314 +trainer/policy/normal/std Std 0.649192 +trainer/policy/normal/std Max 5.96546 +trainer/policy/normal/std Min 0.324167 +trainer/policy/normal/log_std Mean 1.01683 +trainer/policy/normal/log_std Std 0.295702 +trainer/policy/normal/log_std Max 1.78599 +trainer/policy/normal/log_std Min -1.1265 +eval/num steps total 580385 +eval/num paths total 584 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.142496 +eval/Actions Std 0.882204 +eval/Actions Max 0.999981 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82406 +time/logging (s) 0.00423729 +time/sampling batch (s) 0.278168 +time/saving (s) 0.00434565 +time/training (s) 6.06517 +time/epoch (s) 9.17597 +time/total (s) 5981.86 +Epoch -417 +---------------------------------- --------------- +2022-05-10 14:50:35.307002 PDT | [0] Epoch -416 finished +---------------------------------- --------------- +epoch -416 +replay_buffer/size 999033 +trainer/num train calls 585000 +trainer/Policy Loss -19.3656 +trainer/Log Pis Mean 24.232 +trainer/Log Pis Std 13.0814 +trainer/Log Pis Max 63.2096 +trainer/Log Pis Min -9.3167 +trainer/policy/mean Mean -0.03176 +trainer/policy/mean Std 0.90724 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81327 +trainer/policy/normal/std Std 0.641055 +trainer/policy/normal/std Max 5.1035 +trainer/policy/normal/std Min 0.317578 +trainer/policy/normal/log_std Mean 0.995936 +trainer/policy/normal/log_std Std 0.314331 +trainer/policy/normal/log_std Max 1.62993 +trainer/policy/normal/log_std Min -1.14703 +eval/num steps total 581385 +eval/num paths total 585 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.208158 +eval/Actions Std 0.949826 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75861 +time/logging (s) 0.00423622 +time/sampling batch (s) 0.279368 +time/saving (s) 0.00425892 +time/training (s) 6.74298 +time/epoch (s) 9.78945 +time/total (s) 5991.65 +Epoch -416 +---------------------------------- --------------- +2022-05-10 14:50:45.510187 PDT | [0] Epoch -415 finished +---------------------------------- --------------- +epoch -415 +replay_buffer/size 999033 +trainer/num train calls 586000 +trainer/Policy Loss -19.8825 +trainer/Log Pis Mean 26.0386 +trainer/Log Pis Std 13.697 +trainer/Log Pis Max 76.2633 +trainer/Log Pis Min -2.33175 +trainer/policy/mean Mean -0.0462807 +trainer/policy/mean Std 0.909747 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83638 +trainer/policy/normal/std Std 0.655217 +trainer/policy/normal/std Max 6.23521 +trainer/policy/normal/std Min 0.245783 +trainer/policy/normal/log_std Mean 1.00267 +trainer/policy/normal/log_std Std 0.322538 +trainer/policy/normal/log_std Max 1.83021 +trainer/policy/normal/log_std Min -1.40331 +eval/num steps total 582385 +eval/num paths total 586 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0145818 +eval/Actions Std 0.871226 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65731 +time/logging (s) 0.003737 +time/sampling batch (s) 0.532956 +time/saving (s) 0.00360203 +time/training (s) 6.98243 +time/epoch (s) 10.18 +time/total (s) 6001.83 +Epoch -415 +---------------------------------- --------------- +2022-05-10 14:50:56.210019 PDT | [0] Epoch -414 finished +---------------------------------- --------------- +epoch -414 +replay_buffer/size 999033 +trainer/num train calls 587000 +trainer/Policy Loss -19.4746 +trainer/Log Pis Mean 25.0068 +trainer/Log Pis Std 13.3655 +trainer/Log Pis Max 81.6451 +trainer/Log Pis Min -9.08845 +trainer/policy/mean Mean -0.0122006 +trainer/policy/mean Std 0.906066 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76014 +trainer/policy/normal/std Std 0.664811 +trainer/policy/normal/std Max 5.37005 +trainer/policy/normal/std Min 0.241339 +trainer/policy/normal/log_std Mean 0.971151 +trainer/policy/normal/log_std Std 0.340264 +trainer/policy/normal/log_std Max 1.68084 +trainer/policy/normal/log_std Min -1.42155 +eval/num steps total 583385 +eval/num paths total 587 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115223 +eval/Actions Std 0.915089 +eval/Actions Max 0.999974 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50171 +time/logging (s) 0.00370781 +time/sampling batch (s) 0.279579 +time/saving (s) 0.00344069 +time/training (s) 7.88937 +time/epoch (s) 10.6778 +time/total (s) 6012.52 +Epoch -414 +---------------------------------- --------------- +2022-05-10 14:51:05.967956 PDT | [0] Epoch -413 finished +---------------------------------- --------------- +epoch -413 +replay_buffer/size 999033 +trainer/num train calls 588000 +trainer/Policy Loss -19.6708 +trainer/Log Pis Mean 23.66 +trainer/Log Pis Std 12.9776 +trainer/Log Pis Max 59.968 +trainer/Log Pis Min -14.1593 +trainer/policy/mean Mean -0.0288337 +trainer/policy/mean Std 0.905412 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77792 +trainer/policy/normal/std Std 0.653007 +trainer/policy/normal/std Max 6.3843 +trainer/policy/normal/std Min 0.307764 +trainer/policy/normal/log_std Mean 0.981261 +trainer/policy/normal/log_std Std 0.322476 +trainer/policy/normal/log_std Max 1.85384 +trainer/policy/normal/log_std Min -1.17842 +eval/num steps total 584385 +eval/num paths total 588 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114513 +eval/Actions Std 0.913858 +eval/Actions Max 0.999982 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60137 +time/logging (s) 0.00416307 +time/sampling batch (s) 0.281417 +time/saving (s) 0.00411375 +time/training (s) 6.84531 +time/epoch (s) 9.73637 +time/total (s) 6022.25 +Epoch -413 +---------------------------------- --------------- +2022-05-10 14:51:17.413809 PDT | [0] Epoch -412 finished +---------------------------------- --------------- +epoch -412 +replay_buffer/size 999033 +trainer/num train calls 589000 +trainer/Policy Loss -20.2229 +trainer/Log Pis Mean 24.6888 +trainer/Log Pis Std 12.6818 +trainer/Log Pis Max 63.0844 +trainer/Log Pis Min -6.57936 +trainer/policy/mean Mean -0.0544483 +trainer/policy/mean Std 0.905272 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83777 +trainer/policy/normal/std Std 0.637257 +trainer/policy/normal/std Max 5.45465 +trainer/policy/normal/std Min 0.320075 +trainer/policy/normal/log_std Mean 1.00718 +trainer/policy/normal/log_std Std 0.300683 +trainer/policy/normal/log_std Max 1.69647 +trainer/policy/normal/log_std Min -1.1392 +eval/num steps total 585385 +eval/num paths total 589 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.13328 +eval/Actions Std 0.891019 +eval/Actions Max 0.999998 +eval/Actions Min -0.99994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81739 +time/logging (s) 0.00389497 +time/sampling batch (s) 0.285948 +time/saving (s) 0.00380394 +time/training (s) 8.3115 +time/epoch (s) 11.4225 +time/total (s) 6033.68 +Epoch -412 +---------------------------------- --------------- +2022-05-10 14:51:27.812747 PDT | [0] Epoch -411 finished +---------------------------------- --------------- +epoch -411 +replay_buffer/size 999033 +trainer/num train calls 590000 +trainer/Policy Loss -21.1983 +trainer/Log Pis Mean 24.2495 +trainer/Log Pis Std 13.1105 +trainer/Log Pis Max 68.5334 +trainer/Log Pis Min -9.1375 +trainer/policy/mean Mean -0.0439669 +trainer/policy/mean Std 0.908009 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.76732 +trainer/policy/normal/std Std 0.620294 +trainer/policy/normal/std Max 6.10067 +trainer/policy/normal/std Min 0.316243 +trainer/policy/normal/log_std Mean 0.982169 +trainer/policy/normal/log_std Std 0.300095 +trainer/policy/normal/log_std Max 1.8084 +trainer/policy/normal/log_std Min -1.15124 +eval/num steps total 586385 +eval/num paths total 590 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0911678 +eval/Actions Std 0.918042 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58035 +time/logging (s) 0.0036506 +time/sampling batch (s) 0.28756 +time/saving (s) 0.00339554 +time/training (s) 7.50091 +time/epoch (s) 10.3759 +time/total (s) 6044.06 +Epoch -411 +---------------------------------- --------------- +2022-05-10 14:51:37.960748 PDT | [0] Epoch -410 finished +---------------------------------- --------------- +epoch -410 +replay_buffer/size 999033 +trainer/num train calls 591000 +trainer/Policy Loss -21.6746 +trainer/Log Pis Mean 24.6123 +trainer/Log Pis Std 13.1433 +trainer/Log Pis Max 61.2661 +trainer/Log Pis Min -6.00374 +trainer/policy/mean Mean -0.0468711 +trainer/policy/mean Std 0.905309 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80936 +trainer/policy/normal/std Std 0.667026 +trainer/policy/normal/std Max 5.93785 +trainer/policy/normal/std Min 0.335461 +trainer/policy/normal/log_std Mean 0.992125 +trainer/policy/normal/log_std Std 0.323363 +trainer/policy/normal/log_std Max 1.78135 +trainer/policy/normal/log_std Min -1.09225 +eval/num steps total 587385 +eval/num paths total 591 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0627008 +eval/Actions Std 0.916277 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61927 +time/logging (s) 0.00423538 +time/sampling batch (s) 0.28486 +time/saving (s) 0.00438425 +time/training (s) 7.21361 +time/epoch (s) 10.1264 +time/total (s) 6054.19 +Epoch -410 +---------------------------------- --------------- +2022-05-10 14:51:49.169644 PDT | [0] Epoch -409 finished +---------------------------------- --------------- +epoch -409 +replay_buffer/size 999033 +trainer/num train calls 592000 +trainer/Policy Loss -18.4367 +trainer/Log Pis Mean 23.5071 +trainer/Log Pis Std 13.0745 +trainer/Log Pis Max 62.3886 +trainer/Log Pis Min -8.91021 +trainer/policy/mean Mean -0.0339161 +trainer/policy/mean Std 0.899938 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.75499 +trainer/policy/normal/std Std 0.664816 +trainer/policy/normal/std Max 5.67778 +trainer/policy/normal/std Min 0.287407 +trainer/policy/normal/log_std Mean 0.971147 +trainer/policy/normal/log_std Std 0.329552 +trainer/policy/normal/log_std Max 1.73656 +trainer/policy/normal/log_std Min -1.24686 +eval/num steps total 588385 +eval/num paths total 592 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0603078 +eval/Actions Std 0.856971 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62519 +time/logging (s) 0.00406658 +time/sampling batch (s) 0.534966 +time/saving (s) 0.00413191 +time/training (s) 8.01721 +time/epoch (s) 11.1856 +time/total (s) 6065.38 +Epoch -409 +---------------------------------- --------------- +2022-05-10 14:51:59.773568 PDT | [0] Epoch -408 finished +---------------------------------- --------------- +epoch -408 +replay_buffer/size 999033 +trainer/num train calls 593000 +trainer/Policy Loss -20.5631 +trainer/Log Pis Mean 23.7229 +trainer/Log Pis Std 13.3741 +trainer/Log Pis Max 62.0177 +trainer/Log Pis Min -6.25358 +trainer/policy/mean Mean -0.0427144 +trainer/policy/mean Std 0.904406 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8227 +trainer/policy/normal/std Std 0.661212 +trainer/policy/normal/std Max 5.56401 +trainer/policy/normal/std Min 0.340966 +trainer/policy/normal/log_std Mean 0.996638 +trainer/policy/normal/log_std Std 0.32705 +trainer/policy/normal/log_std Max 1.71632 +trainer/policy/normal/log_std Min -1.07597 +eval/num steps total 589385 +eval/num paths total 593 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0993046 +eval/Actions Std 0.894165 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62181 +time/logging (s) 0.00397127 +time/sampling batch (s) 0.285517 +time/saving (s) 0.00410339 +time/training (s) 7.66569 +time/epoch (s) 10.5811 +time/total (s) 6075.96 +Epoch -408 +---------------------------------- --------------- +2022-05-10 14:52:09.638021 PDT | [0] Epoch -407 finished +---------------------------------- --------------- +epoch -407 +replay_buffer/size 999033 +trainer/num train calls 594000 +trainer/Policy Loss -18.7053 +trainer/Log Pis Mean 23.6408 +trainer/Log Pis Std 12.9078 +trainer/Log Pis Max 64.4481 +trainer/Log Pis Min -9.49757 +trainer/policy/mean Mean -0.022132 +trainer/policy/mean Std 0.907658 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.76966 +trainer/policy/normal/std Std 0.646084 +trainer/policy/normal/std Max 6.42123 +trainer/policy/normal/std Min 0.286821 +trainer/policy/normal/log_std Mean 0.979421 +trainer/policy/normal/log_std Std 0.316254 +trainer/policy/normal/log_std Max 1.85961 +trainer/policy/normal/log_std Min -1.2489 +eval/num steps total 590385 +eval/num paths total 594 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00996631 +eval/Actions Std 0.904814 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57326 +time/logging (s) 0.00417615 +time/sampling batch (s) 0.28063 +time/saving (s) 0.0043027 +time/training (s) 6.97985 +time/epoch (s) 9.84222 +time/total (s) 6085.81 +Epoch -407 +---------------------------------- --------------- +2022-05-10 14:52:20.326406 PDT | [0] Epoch -406 finished +---------------------------------- --------------- +epoch -406 +replay_buffer/size 999033 +trainer/num train calls 595000 +trainer/Policy Loss -19.0663 +trainer/Log Pis Mean 23.3087 +trainer/Log Pis Std 12.4882 +trainer/Log Pis Max 75.5568 +trainer/Log Pis Min -5.77633 +trainer/policy/mean Mean -0.0348125 +trainer/policy/mean Std 0.902656 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.74383 +trainer/policy/normal/std Std 0.612495 +trainer/policy/normal/std Max 5.40368 +trainer/policy/normal/std Min 0.359607 +trainer/policy/normal/log_std Mean 0.973579 +trainer/policy/normal/log_std Std 0.300577 +trainer/policy/normal/log_std Max 1.68708 +trainer/policy/normal/log_std Min -1.02274 +eval/num steps total 591385 +eval/num paths total 595 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.161365 +eval/Actions Std 0.895605 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66155 +time/logging (s) 0.00372512 +time/sampling batch (s) 0.283453 +time/saving (s) 0.00347677 +time/training (s) 7.71299 +time/epoch (s) 10.6652 +time/total (s) 6096.48 +Epoch -406 +---------------------------------- --------------- +2022-05-10 14:52:31.171980 PDT | [0] Epoch -405 finished +---------------------------------- --------------- +epoch -405 +replay_buffer/size 999033 +trainer/num train calls 596000 +trainer/Policy Loss -20.0919 +trainer/Log Pis Mean 24.2465 +trainer/Log Pis Std 12.8341 +trainer/Log Pis Max 69.3867 +trainer/Log Pis Min -9.74546 +trainer/policy/mean Mean -0.0317802 +trainer/policy/mean Std 0.902334 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.69282 +trainer/policy/normal/std Std 0.652256 +trainer/policy/normal/std Max 4.9685 +trainer/policy/normal/std Min 0.273359 +trainer/policy/normal/log_std Mean 0.946485 +trainer/policy/normal/log_std Std 0.338688 +trainer/policy/normal/log_std Max 1.60312 +trainer/policy/normal/log_std Min -1.29697 +eval/num steps total 592385 +eval/num paths total 596 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0194164 +eval/Actions Std 0.924125 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61345 +time/logging (s) 0.00376534 +time/sampling batch (s) 0.786711 +time/saving (s) 0.00345984 +time/training (s) 7.4157 +time/epoch (s) 10.8231 +time/total (s) 6107.3 +Epoch -405 +---------------------------------- --------------- +2022-05-10 14:52:40.773840 PDT | [0] Epoch -404 finished +---------------------------------- --------------- +epoch -404 +replay_buffer/size 999033 +trainer/num train calls 597000 +trainer/Policy Loss -20.5142 +trainer/Log Pis Mean 25.062 +trainer/Log Pis Std 13.6074 +trainer/Log Pis Max 65.4003 +trainer/Log Pis Min -4.22734 +trainer/policy/mean Mean -0.0351927 +trainer/policy/mean Std 0.910791 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8143 +trainer/policy/normal/std Std 0.654142 +trainer/policy/normal/std Max 6.27784 +trainer/policy/normal/std Min 0.333777 +trainer/policy/normal/log_std Mean 0.995079 +trainer/policy/normal/log_std Std 0.318955 +trainer/policy/normal/log_std Max 1.83703 +trainer/policy/normal/log_std Min -1.09728 +eval/num steps total 593385 +eval/num paths total 597 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.35184 +eval/Actions Std 0.857184 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65745 +time/logging (s) 0.00397125 +time/sampling batch (s) 0.28379 +time/saving (s) 0.0042796 +time/training (s) 6.63006 +time/epoch (s) 9.57955 +time/total (s) 6116.88 +Epoch -404 +---------------------------------- --------------- +2022-05-10 14:52:50.984802 PDT | [0] Epoch -403 finished +---------------------------------- --------------- +epoch -403 +replay_buffer/size 999033 +trainer/num train calls 598000 +trainer/Policy Loss -19.4989 +trainer/Log Pis Mean 24.4424 +trainer/Log Pis Std 13.2787 +trainer/Log Pis Max 70.1677 +trainer/Log Pis Min -5.76194 +trainer/policy/mean Mean -0.0555616 +trainer/policy/mean Std 0.906863 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80122 +trainer/policy/normal/std Std 0.655967 +trainer/policy/normal/std Max 6.20138 +trainer/policy/normal/std Min 0.324427 +trainer/policy/normal/log_std Mean 0.990302 +trainer/policy/normal/log_std Std 0.319206 +trainer/policy/normal/log_std Max 1.82477 +trainer/policy/normal/log_std Min -1.1257 +eval/num steps total 594385 +eval/num paths total 598 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0203173 +eval/Actions Std 0.86232 +eval/Actions Max 0.999999 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56909 +time/logging (s) 0.00368929 +time/sampling batch (s) 0.282548 +time/saving (s) 0.00349613 +time/training (s) 7.32927 +time/epoch (s) 10.1881 +time/total (s) 6127.08 +Epoch -403 +---------------------------------- --------------- +2022-05-10 14:53:01.271976 PDT | [0] Epoch -402 finished +---------------------------------- --------------- +epoch -402 +replay_buffer/size 999033 +trainer/num train calls 599000 +trainer/Policy Loss -19.9254 +trainer/Log Pis Mean 24.011 +trainer/Log Pis Std 12.0937 +trainer/Log Pis Max 66.5912 +trainer/Log Pis Min -5.84531 +trainer/policy/mean Mean -0.0198685 +trainer/policy/mean Std 0.903413 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.81086 +trainer/policy/normal/std Std 0.650919 +trainer/policy/normal/std Max 6.69456 +trainer/policy/normal/std Min 0.281538 +trainer/policy/normal/log_std Mean 0.995514 +trainer/policy/normal/log_std Std 0.310347 +trainer/policy/normal/log_std Max 1.90129 +trainer/policy/normal/log_std Min -1.26749 +eval/num steps total 595385 +eval/num paths total 599 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00188423 +eval/Actions Std 0.912464 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69907 +time/logging (s) 0.00381593 +time/sampling batch (s) 0.281633 +time/saving (s) 0.00347553 +time/training (s) 7.27706 +time/epoch (s) 10.2651 +time/total (s) 6137.34 +Epoch -402 +---------------------------------- --------------- +2022-05-10 14:53:11.273831 PDT | [0] Epoch -401 finished +---------------------------------- --------------- +epoch -401 +replay_buffer/size 999033 +trainer/num train calls 600000 +trainer/Policy Loss -18.845 +trainer/Log Pis Mean 23.8748 +trainer/Log Pis Std 12.9899 +trainer/Log Pis Max 62.6597 +trainer/Log Pis Min -11.1048 +trainer/policy/mean Mean -0.0634202 +trainer/policy/mean Std 0.901667 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77892 +trainer/policy/normal/std Std 0.675493 +trainer/policy/normal/std Max 5.85619 +trainer/policy/normal/std Min 0.258416 +trainer/policy/normal/log_std Mean 0.978032 +trainer/policy/normal/log_std Std 0.337759 +trainer/policy/normal/log_std Max 1.7675 +trainer/policy/normal/log_std Min -1.35318 +eval/num steps total 596385 +eval/num paths total 600 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0591548 +eval/Actions Std 0.897477 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34174 +time/logging (s) 0.00374047 +time/sampling batch (s) 0.281885 +time/saving (s) 0.00637069 +time/training (s) 7.34575 +time/epoch (s) 9.97949 +time/total (s) 6147.33 +Epoch -401 +---------------------------------- --------------- +2022-05-10 14:53:21.085222 PDT | [0] Epoch -400 finished +---------------------------------- --------------- +epoch -400 +replay_buffer/size 999033 +trainer/num train calls 601000 +trainer/Policy Loss -20.5375 +trainer/Log Pis Mean 24.0225 +trainer/Log Pis Std 13.4251 +trainer/Log Pis Max 65.8906 +trainer/Log Pis Min -9.30169 +trainer/policy/mean Mean -0.0431795 +trainer/policy/mean Std 0.90921 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.75536 +trainer/policy/normal/std Std 0.634423 +trainer/policy/normal/std Max 6.25135 +trainer/policy/normal/std Min 0.284408 +trainer/policy/normal/log_std Mean 0.976205 +trainer/policy/normal/log_std Std 0.305995 +trainer/policy/normal/log_std Max 1.8328 +trainer/policy/normal/log_std Min -1.25735 +eval/num steps total 597385 +eval/num paths total 601 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0843378 +eval/Actions Std 0.929793 +eval/Actions Max 0.999992 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56437 +time/logging (s) 0.00367863 +time/sampling batch (s) 0.28183 +time/saving (s) 0.00349578 +time/training (s) 6.93586 +time/epoch (s) 9.78923 +time/total (s) 6157.12 +Epoch -400 +---------------------------------- --------------- +2022-05-10 14:53:31.235810 PDT | [0] Epoch -399 finished +---------------------------------- --------------- +epoch -399 +replay_buffer/size 999033 +trainer/num train calls 602000 +trainer/Policy Loss -18.9248 +trainer/Log Pis Mean 23.5234 +trainer/Log Pis Std 12.9712 +trainer/Log Pis Max 63.6489 +trainer/Log Pis Min -10.9546 +trainer/policy/mean Mean -0.0438129 +trainer/policy/mean Std 0.903553 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79876 +trainer/policy/normal/std Std 0.669844 +trainer/policy/normal/std Max 6.17173 +trainer/policy/normal/std Min 0.340212 +trainer/policy/normal/log_std Mean 0.988358 +trainer/policy/normal/log_std Std 0.320825 +trainer/policy/normal/log_std Max 1.81998 +trainer/policy/normal/log_std Min -1.07818 +eval/num steps total 598385 +eval/num paths total 602 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.292742 +eval/Actions Std 0.882367 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59299 +time/logging (s) 0.00385208 +time/sampling batch (s) 0.533645 +time/saving (s) 0.00416884 +time/training (s) 6.99385 +time/epoch (s) 10.1285 +time/total (s) 6167.25 +Epoch -399 +---------------------------------- --------------- +2022-05-10 14:53:41.949019 PDT | [0] Epoch -398 finished +---------------------------------- --------------- +epoch -398 +replay_buffer/size 999033 +trainer/num train calls 603000 +trainer/Policy Loss -20.7379 +trainer/Log Pis Mean 24.1401 +trainer/Log Pis Std 13.2386 +trainer/Log Pis Max 74.2741 +trainer/Log Pis Min -9.51051 +trainer/policy/mean Mean -0.0357684 +trainer/policy/mean Std 0.909139 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77398 +trainer/policy/normal/std Std 0.621803 +trainer/policy/normal/std Max 5.67806 +trainer/policy/normal/std Min 0.305772 +trainer/policy/normal/log_std Mean 0.983715 +trainer/policy/normal/log_std Std 0.306426 +trainer/policy/normal/log_std Max 1.73661 +trainer/policy/normal/log_std Min -1.18492 +eval/num steps total 599385 +eval/num paths total 603 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107774 +eval/Actions Std 0.917678 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59216 +time/logging (s) 0.00369592 +time/sampling batch (s) 0.534006 +time/saving (s) 0.00346561 +time/training (s) 7.55694 +time/epoch (s) 10.6903 +time/total (s) 6177.94 +Epoch -398 +---------------------------------- --------------- +2022-05-10 14:53:52.860005 PDT | [0] Epoch -397 finished +---------------------------------- --------------- +epoch -397 +replay_buffer/size 999033 +trainer/num train calls 604000 +trainer/Policy Loss -19.6328 +trainer/Log Pis Mean 24.5832 +trainer/Log Pis Std 12.7741 +trainer/Log Pis Max 81.1885 +trainer/Log Pis Min -10.8502 +trainer/policy/mean Mean -0.0458388 +trainer/policy/mean Std 0.907414 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79099 +trainer/policy/normal/std Std 0.640406 +trainer/policy/normal/std Max 5.83013 +trainer/policy/normal/std Min 0.291368 +trainer/policy/normal/log_std Mean 0.988965 +trainer/policy/normal/log_std Std 0.307979 +trainer/policy/normal/log_std Max 1.76304 +trainer/policy/normal/log_std Min -1.23317 +eval/num steps total 600385 +eval/num paths total 604 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00199988 +eval/Actions Std 0.952637 +eval/Actions Max 1 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68495 +time/logging (s) 0.00379559 +time/sampling batch (s) 0.28323 +time/saving (s) 0.0034726 +time/training (s) 7.91313 +time/epoch (s) 10.8886 +time/total (s) 6188.83 +Epoch -397 +---------------------------------- --------------- +2022-05-10 14:54:02.846233 PDT | [0] Epoch -396 finished +---------------------------------- --------------- +epoch -396 +replay_buffer/size 999033 +trainer/num train calls 605000 +trainer/Policy Loss -19.7385 +trainer/Log Pis Mean 24.0016 +trainer/Log Pis Std 12.9324 +trainer/Log Pis Max 67.4361 +trainer/Log Pis Min -11.4678 +trainer/policy/mean Mean -0.0290003 +trainer/policy/mean Std 0.90847 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84433 +trainer/policy/normal/std Std 0.644824 +trainer/policy/normal/std Max 5.27207 +trainer/policy/normal/std Min 0.274558 +trainer/policy/normal/log_std Mean 1.00803 +trainer/policy/normal/log_std Std 0.308544 +trainer/policy/normal/log_std Max 1.66242 +trainer/policy/normal/log_std Min -1.29259 +eval/num steps total 601385 +eval/num paths total 605 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104621 +eval/Actions Std 0.886033 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52668 +time/logging (s) 0.00384665 +time/sampling batch (s) 0.529706 +time/saving (s) 0.00349624 +time/training (s) 6.90062 +time/epoch (s) 9.96435 +time/total (s) 6198.8 +Epoch -396 +---------------------------------- --------------- +2022-05-10 14:54:13.282115 PDT | [0] Epoch -395 finished +---------------------------------- --------------- +epoch -395 +replay_buffer/size 999033 +trainer/num train calls 606000 +trainer/Policy Loss -21.2641 +trainer/Log Pis Mean 24.179 +trainer/Log Pis Std 13.1685 +trainer/Log Pis Max 73.496 +trainer/Log Pis Min -8.39029 +trainer/policy/mean Mean -0.0361607 +trainer/policy/mean Std 0.904564 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80087 +trainer/policy/normal/std Std 0.624467 +trainer/policy/normal/std Max 7.3389 +trainer/policy/normal/std Min 0.341153 +trainer/policy/normal/log_std Mean 0.996247 +trainer/policy/normal/log_std Std 0.287928 +trainer/policy/normal/log_std Max 1.99319 +trainer/policy/normal/log_std Min -1.07543 +eval/num steps total 602385 +eval/num paths total 606 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101833 +eval/Actions Std 0.904891 +eval/Actions Max 0.999986 +eval/Actions Min -0.999973 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69634 +time/logging (s) 0.00442959 +time/sampling batch (s) 0.283735 +time/saving (s) 0.00393405 +time/training (s) 7.42565 +time/epoch (s) 10.4141 +time/total (s) 6209.22 +Epoch -395 +---------------------------------- --------------- +2022-05-10 14:54:23.984156 PDT | [0] Epoch -394 finished +---------------------------------- --------------- +epoch -394 +replay_buffer/size 999033 +trainer/num train calls 607000 +trainer/Policy Loss -19.3927 +trainer/Log Pis Mean 24.7828 +trainer/Log Pis Std 13.1984 +trainer/Log Pis Max 65.1377 +trainer/Log Pis Min -9.38434 +trainer/policy/mean Mean -0.0422713 +trainer/policy/mean Std 0.906399 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78705 +trainer/policy/normal/std Std 0.663142 +trainer/policy/normal/std Max 5.50778 +trainer/policy/normal/std Min 0.291725 +trainer/policy/normal/log_std Mean 0.983746 +trainer/policy/normal/log_std Std 0.325164 +trainer/policy/normal/log_std Max 1.70616 +trainer/policy/normal/log_std Min -1.23194 +eval/num steps total 603385 +eval/num paths total 607 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0166286 +eval/Actions Std 0.90641 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65549 +time/logging (s) 0.00420402 +time/sampling batch (s) 0.292009 +time/saving (s) 0.00428603 +time/training (s) 7.72214 +time/epoch (s) 10.6781 +time/total (s) 6219.9 +Epoch -394 +---------------------------------- --------------- +2022-05-10 14:54:33.999785 PDT | [0] Epoch -393 finished +---------------------------------- --------------- +epoch -393 +replay_buffer/size 999033 +trainer/num train calls 608000 +trainer/Policy Loss -19.6489 +trainer/Log Pis Mean 25.3058 +trainer/Log Pis Std 13.2897 +trainer/Log Pis Max 71.7896 +trainer/Log Pis Min -10.114 +trainer/policy/mean Mean -0.0427074 +trainer/policy/mean Std 0.912928 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81257 +trainer/policy/normal/std Std 0.646379 +trainer/policy/normal/std Max 6.36147 +trainer/policy/normal/std Min 0.224737 +trainer/policy/normal/log_std Mean 0.996364 +trainer/policy/normal/log_std Std 0.30964 +trainer/policy/normal/log_std Max 1.85026 +trainer/policy/normal/log_std Min -1.49282 +eval/num steps total 604385 +eval/num paths total 608 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0110911 +eval/Actions Std 0.912779 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64825 +time/logging (s) 0.00378138 +time/sampling batch (s) 0.290129 +time/saving (s) 0.00355465 +time/training (s) 7.04629 +time/epoch (s) 9.992 +time/total (s) 6229.9 +Epoch -393 +---------------------------------- --------------- +2022-05-10 14:54:44.183661 PDT | [0] Epoch -392 finished +---------------------------------- --------------- +epoch -392 +replay_buffer/size 999033 +trainer/num train calls 609000 +trainer/Policy Loss -19.4854 +trainer/Log Pis Mean 24.7601 +trainer/Log Pis Std 12.95 +trainer/Log Pis Max 70.2022 +trainer/Log Pis Min -9.1016 +trainer/policy/mean Mean -0.0372947 +trainer/policy/mean Std 0.905475 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.7968 +trainer/policy/normal/std Std 0.662611 +trainer/policy/normal/std Max 5.35796 +trainer/policy/normal/std Min 0.247545 +trainer/policy/normal/log_std Mean 0.987422 +trainer/policy/normal/log_std Std 0.324877 +trainer/policy/normal/log_std Max 1.67858 +trainer/policy/normal/log_std Min -1.39616 +eval/num steps total 605385 +eval/num paths total 609 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0286624 +eval/Actions Std 0.901574 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57179 +time/logging (s) 0.00403826 +time/sampling batch (s) 0.294352 +time/saving (s) 0.0035995 +time/training (s) 7.28721 +time/epoch (s) 10.161 +time/total (s) 6240.06 +Epoch -392 +---------------------------------- --------------- +2022-05-10 14:54:54.647029 PDT | [0] Epoch -391 finished +---------------------------------- --------------- +epoch -391 +replay_buffer/size 999033 +trainer/num train calls 610000 +trainer/Policy Loss -21.0026 +trainer/Log Pis Mean 24.6842 +trainer/Log Pis Std 14.05 +trainer/Log Pis Max 70.097 +trainer/Log Pis Min -8.60734 +trainer/policy/mean Mean -0.037186 +trainer/policy/mean Std 0.905523 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77922 +trainer/policy/normal/std Std 0.662479 +trainer/policy/normal/std Max 5.28827 +trainer/policy/normal/std Min 0.252785 +trainer/policy/normal/log_std Mean 0.979595 +trainer/policy/normal/log_std Std 0.332832 +trainer/policy/normal/log_std Max 1.66549 +trainer/policy/normal/log_std Min -1.37522 +eval/num steps total 606385 +eval/num paths total 610 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.452562 +eval/Actions Std 0.80728 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54141 +time/logging (s) 0.00371757 +time/sampling batch (s) 0.287347 +time/saving (s) 0.00357499 +time/training (s) 7.60415 +time/epoch (s) 10.4402 +time/total (s) 6250.5 +Epoch -391 +---------------------------------- --------------- +2022-05-10 14:55:05.003748 PDT | [0] Epoch -390 finished +---------------------------------- --------------- +epoch -390 +replay_buffer/size 999033 +trainer/num train calls 611000 +trainer/Policy Loss -19.6398 +trainer/Log Pis Mean 23.2557 +trainer/Log Pis Std 13.3209 +trainer/Log Pis Max 66.4482 +trainer/Log Pis Min -7.93927 +trainer/policy/mean Mean -0.0367366 +trainer/policy/mean Std 0.906746 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79331 +trainer/policy/normal/std Std 0.657152 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.31577 +trainer/policy/normal/log_std Mean 0.986334 +trainer/policy/normal/log_std Std 0.324926 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.15274 +eval/num steps total 607385 +eval/num paths total 611 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.015202 +eval/Actions Std 0.911707 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49822 +time/logging (s) 0.00377388 +time/sampling batch (s) 0.280502 +time/saving (s) 0.00347997 +time/training (s) 7.54867 +time/epoch (s) 10.3346 +time/total (s) 6260.84 +Epoch -390 +---------------------------------- --------------- +2022-05-10 14:55:15.085169 PDT | [0] Epoch -389 finished +---------------------------------- --------------- +epoch -389 +replay_buffer/size 999033 +trainer/num train calls 612000 +trainer/Policy Loss -18.7197 +trainer/Log Pis Mean 23.9835 +trainer/Log Pis Std 12.6267 +trainer/Log Pis Max 77.5386 +trainer/Log Pis Min -5.51045 +trainer/policy/mean Mean -0.0486657 +trainer/policy/mean Std 0.905194 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74087 +trainer/policy/normal/std Std 0.641963 +trainer/policy/normal/std Max 6.03318 +trainer/policy/normal/std Min 0.336965 +trainer/policy/normal/log_std Mean 0.96795 +trainer/policy/normal/log_std Std 0.322229 +trainer/policy/normal/log_std Max 1.79727 +trainer/policy/normal/log_std Min -1.08778 +eval/num steps total 608385 +eval/num paths total 612 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0459877 +eval/Actions Std 0.908109 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58769 +time/logging (s) 0.00421314 +time/sampling batch (s) 0.278065 +time/saving (s) 0.00412825 +time/training (s) 7.18565 +time/epoch (s) 10.0597 +time/total (s) 6270.9 +Epoch -389 +---------------------------------- --------------- +2022-05-10 14:55:24.732762 PDT | [0] Epoch -388 finished +---------------------------------- --------------- +epoch -388 +replay_buffer/size 999033 +trainer/num train calls 613000 +trainer/Policy Loss -19.386 +trainer/Log Pis Mean 23.9061 +trainer/Log Pis Std 13.1179 +trainer/Log Pis Max 74.4303 +trainer/Log Pis Min -8.01556 +trainer/policy/mean Mean -0.0420675 +trainer/policy/mean Std 0.902046 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76374 +trainer/policy/normal/std Std 0.639287 +trainer/policy/normal/std Max 4.96772 +trainer/policy/normal/std Min 0.264537 +trainer/policy/normal/log_std Mean 0.97699 +trainer/policy/normal/log_std Std 0.319819 +trainer/policy/normal/log_std Max 1.60296 +trainer/policy/normal/log_std Min -1.32977 +eval/num steps total 609385 +eval/num paths total 613 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.125602 +eval/Actions Std 0.904947 +eval/Actions Max 0.999971 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54557 +time/logging (s) 0.00377557 +time/sampling batch (s) 0.28195 +time/saving (s) 0.00357723 +time/training (s) 6.78956 +time/epoch (s) 9.62443 +time/total (s) 6280.53 +Epoch -388 +---------------------------------- --------------- +2022-05-10 14:55:33.873315 PDT | [0] Epoch -387 finished +---------------------------------- --------------- +epoch -387 +replay_buffer/size 999033 +trainer/num train calls 614000 +trainer/Policy Loss -18.947 +trainer/Log Pis Mean 24.5028 +trainer/Log Pis Std 12.9236 +trainer/Log Pis Max 61.6275 +trainer/Log Pis Min -9.74856 +trainer/policy/mean Mean -0.0351127 +trainer/policy/mean Std 0.907407 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.79896 +trainer/policy/normal/std Std 0.668812 +trainer/policy/normal/std Max 7.08129 +trainer/policy/normal/std Min 0.289988 +trainer/policy/normal/log_std Mean 0.987716 +trainer/policy/normal/log_std Std 0.326952 +trainer/policy/normal/log_std Max 1.95746 +trainer/policy/normal/log_std Min -1.23792 +eval/num steps total 610385 +eval/num paths total 614 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114811 +eval/Actions Std 0.912989 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5844 +time/logging (s) 0.00373073 +time/sampling batch (s) 0.284408 +time/saving (s) 0.00347778 +time/training (s) 6.24216 +time/epoch (s) 9.11817 +time/total (s) 6289.65 +Epoch -387 +---------------------------------- --------------- +2022-05-10 14:55:43.114033 PDT | [0] Epoch -386 finished +---------------------------------- --------------- +epoch -386 +replay_buffer/size 999033 +trainer/num train calls 615000 +trainer/Policy Loss -18.4878 +trainer/Log Pis Mean 24.4212 +trainer/Log Pis Std 12.973 +trainer/Log Pis Max 61.5792 +trainer/Log Pis Min -7.45754 +trainer/policy/mean Mean -0.0282176 +trainer/policy/mean Std 0.905912 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83473 +trainer/policy/normal/std Std 0.66032 +trainer/policy/normal/std Max 6.91545 +trainer/policy/normal/std Min 0.289484 +trainer/policy/normal/log_std Mean 1.00283 +trainer/policy/normal/log_std Std 0.316846 +trainer/policy/normal/log_std Max 1.93376 +trainer/policy/normal/log_std Min -1.23965 +eval/num steps total 611385 +eval/num paths total 615 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167537 +eval/Actions Std 0.913172 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62969 +time/logging (s) 0.00376247 +time/sampling batch (s) 0.285339 +time/saving (s) 0.00349382 +time/training (s) 6.29593 +time/epoch (s) 9.21822 +time/total (s) 6298.87 +Epoch -386 +---------------------------------- --------------- +2022-05-10 14:55:53.139254 PDT | [0] Epoch -385 finished +---------------------------------- --------------- +epoch -385 +replay_buffer/size 999033 +trainer/num train calls 616000 +trainer/Policy Loss -18.4691 +trainer/Log Pis Mean 24.4595 +trainer/Log Pis Std 13.4314 +trainer/Log Pis Max 73.8497 +trainer/Log Pis Min -2.74492 +trainer/policy/mean Mean -0.0368878 +trainer/policy/mean Std 0.907684 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78844 +trainer/policy/normal/std Std 0.660287 +trainer/policy/normal/std Max 5.5782 +trainer/policy/normal/std Min 0.288647 +trainer/policy/normal/log_std Mean 0.982397 +trainer/policy/normal/log_std Std 0.337299 +trainer/policy/normal/log_std Max 1.71887 +trainer/policy/normal/log_std Min -1.24255 +eval/num steps total 612385 +eval/num paths total 616 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.170201 +eval/Actions Std 0.924425 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61321 +time/logging (s) 0.00379924 +time/sampling batch (s) 0.533327 +time/saving (s) 0.00348549 +time/training (s) 6.849 +time/epoch (s) 10.0028 +time/total (s) 6308.88 +Epoch -385 +---------------------------------- --------------- +2022-05-10 14:56:03.154371 PDT | [0] Epoch -384 finished +---------------------------------- --------------- +epoch -384 +replay_buffer/size 999033 +trainer/num train calls 617000 +trainer/Policy Loss -20.0585 +trainer/Log Pis Mean 25.3292 +trainer/Log Pis Std 13.2314 +trainer/Log Pis Max 71.3912 +trainer/Log Pis Min -12.6184 +trainer/policy/mean Mean -0.0543008 +trainer/policy/mean Std 0.909446 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8993 +trainer/policy/normal/std Std 0.652497 +trainer/policy/normal/std Max 6.34716 +trainer/policy/normal/std Min 0.313498 +trainer/policy/normal/log_std Mean 1.02858 +trainer/policy/normal/log_std Std 0.301551 +trainer/policy/normal/log_std Max 1.84801 +trainer/policy/normal/log_std Min -1.15996 +eval/num steps total 613385 +eval/num paths total 617 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121439 +eval/Actions Std 0.895146 +eval/Actions Max 0.999997 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66801 +time/logging (s) 0.00415642 +time/sampling batch (s) 0.532541 +time/saving (s) 0.00415308 +time/training (s) 6.78417 +time/epoch (s) 9.99303 +time/total (s) 6318.87 +Epoch -384 +---------------------------------- --------------- +2022-05-10 14:56:13.290682 PDT | [0] Epoch -383 finished +---------------------------------- --------------- +epoch -383 +replay_buffer/size 999033 +trainer/num train calls 618000 +trainer/Policy Loss -20.1774 +trainer/Log Pis Mean 23.6535 +trainer/Log Pis Std 12.495 +trainer/Log Pis Max 73.0655 +trainer/Log Pis Min -8.47788 +trainer/policy/mean Mean -0.0475651 +trainer/policy/mean Std 0.905996 +trainer/policy/mean Max 0.999962 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78543 +trainer/policy/normal/std Std 0.651663 +trainer/policy/normal/std Max 5.15515 +trainer/policy/normal/std Min 0.355385 +trainer/policy/normal/log_std Mean 0.984112 +trainer/policy/normal/log_std Std 0.321314 +trainer/policy/normal/log_std Max 1.64 +trainer/policy/normal/log_std Min -1.03455 +eval/num steps total 614385 +eval/num paths total 618 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232258 +eval/Actions Std 0.888588 +eval/Actions Max 0.999982 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74946 +time/logging (s) 0.00377625 +time/sampling batch (s) 0.529962 +time/saving (s) 0.00371051 +time/training (s) 6.82643 +time/epoch (s) 10.1133 +time/total (s) 6328.99 +Epoch -383 +---------------------------------- --------------- +2022-05-10 14:56:23.181664 PDT | [0] Epoch -382 finished +---------------------------------- --------------- +epoch -382 +replay_buffer/size 999033 +trainer/num train calls 619000 +trainer/Policy Loss -19.1582 +trainer/Log Pis Mean 22.8762 +trainer/Log Pis Std 13.1962 +trainer/Log Pis Max 63.9484 +trainer/Log Pis Min -13.9036 +trainer/policy/mean Mean -0.0253678 +trainer/policy/mean Std 0.902543 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81761 +trainer/policy/normal/std Std 0.675434 +trainer/policy/normal/std Max 6.40169 +trainer/policy/normal/std Min 0.242942 +trainer/policy/normal/log_std Mean 0.995483 +trainer/policy/normal/log_std Std 0.319765 +trainer/policy/normal/log_std Max 1.85656 +trainer/policy/normal/log_std Min -1.41493 +eval/num steps total 615385 +eval/num paths total 619 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0710162 +eval/Actions Std 0.899118 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62301 +time/logging (s) 0.00373025 +time/sampling batch (s) 0.291977 +time/saving (s) 0.00339516 +time/training (s) 6.94616 +time/epoch (s) 9.86827 +time/total (s) 6338.86 +Epoch -382 +---------------------------------- --------------- +2022-05-10 14:56:33.756083 PDT | [0] Epoch -381 finished +---------------------------------- --------------- +epoch -381 +replay_buffer/size 999033 +trainer/num train calls 620000 +trainer/Policy Loss -19.1246 +trainer/Log Pis Mean 24.0675 +trainer/Log Pis Std 13.6422 +trainer/Log Pis Max 78.35 +trainer/Log Pis Min -6.06315 +trainer/policy/mean Mean -0.0323002 +trainer/policy/mean Std 0.906441 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79092 +trainer/policy/normal/std Std 0.668176 +trainer/policy/normal/std Max 5.52882 +trainer/policy/normal/std Min 0.312331 +trainer/policy/normal/log_std Mean 0.984313 +trainer/policy/normal/log_std Std 0.329432 +trainer/policy/normal/log_std Max 1.70997 +trainer/policy/normal/log_std Min -1.16369 +eval/num steps total 616385 +eval/num paths total 620 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0680537 +eval/Actions Std 0.917971 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57998 +time/logging (s) 0.00366245 +time/sampling batch (s) 0.529776 +time/saving (s) 0.00357971 +time/training (s) 7.43548 +time/epoch (s) 10.5525 +time/total (s) 6349.42 +Epoch -381 +---------------------------------- --------------- +2022-05-10 14:56:43.565602 PDT | [0] Epoch -380 finished +---------------------------------- --------------- +epoch -380 +replay_buffer/size 999033 +trainer/num train calls 621000 +trainer/Policy Loss -19.0954 +trainer/Log Pis Mean 24.8526 +trainer/Log Pis Std 13.4529 +trainer/Log Pis Max 71.2676 +trainer/Log Pis Min -4.14885 +trainer/policy/mean Mean -0.0293633 +trainer/policy/mean Std 0.90917 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82552 +trainer/policy/normal/std Std 0.651681 +trainer/policy/normal/std Max 6.02354 +trainer/policy/normal/std Min 0.26401 +trainer/policy/normal/log_std Mean 1.00047 +trainer/policy/normal/log_std Std 0.312666 +trainer/policy/normal/log_std Max 1.79567 +trainer/policy/normal/log_std Min -1.33177 +eval/num steps total 617385 +eval/num paths total 621 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0985185 +eval/Actions Std 0.91544 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58762 +time/logging (s) 0.0037114 +time/sampling batch (s) 0.52743 +time/saving (s) 0.00340925 +time/training (s) 6.66557 +time/epoch (s) 9.78773 +time/total (s) 6359.21 +Epoch -380 +---------------------------------- --------------- +2022-05-10 14:56:52.608793 PDT | [0] Epoch -379 finished +---------------------------------- --------------- +epoch -379 +replay_buffer/size 999033 +trainer/num train calls 622000 +trainer/Policy Loss -20.0453 +trainer/Log Pis Mean 24.8924 +trainer/Log Pis Std 13.0091 +trainer/Log Pis Max 66.1481 +trainer/Log Pis Min -4.93793 +trainer/policy/mean Mean -0.0316932 +trainer/policy/mean Std 0.908557 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.82165 +trainer/policy/normal/std Std 0.640091 +trainer/policy/normal/std Max 5.22236 +trainer/policy/normal/std Min 0.311423 +trainer/policy/normal/log_std Mean 1.00005 +trainer/policy/normal/log_std Std 0.308275 +trainer/policy/normal/log_std Max 1.65295 +trainer/policy/normal/log_std Min -1.1666 +eval/num steps total 618385 +eval/num paths total 622 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0102144 +eval/Actions Std 0.921613 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48744 +time/logging (s) 0.00374672 +time/sampling batch (s) 0.278311 +time/saving (s) 0.00346543 +time/training (s) 6.24828 +time/epoch (s) 9.02124 +time/total (s) 6368.23 +Epoch -379 +---------------------------------- --------------- +2022-05-10 14:57:01.751565 PDT | [0] Epoch -378 finished +---------------------------------- --------------- +epoch -378 +replay_buffer/size 999033 +trainer/num train calls 623000 +trainer/Policy Loss -20.6182 +trainer/Log Pis Mean 25.5692 +trainer/Log Pis Std 12.8274 +trainer/Log Pis Max 73.9814 +trainer/Log Pis Min -8.26808 +trainer/policy/mean Mean -0.0340337 +trainer/policy/mean Std 0.907717 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81314 +trainer/policy/normal/std Std 0.643575 +trainer/policy/normal/std Max 6.09983 +trainer/policy/normal/std Min 0.301416 +trainer/policy/normal/log_std Mean 0.995445 +trainer/policy/normal/log_std Std 0.317881 +trainer/policy/normal/log_std Max 1.80826 +trainer/policy/normal/log_std Min -1.19927 +eval/num steps total 619385 +eval/num paths total 623 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.019034 +eval/Actions Std 0.907987 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47251 +time/logging (s) 0.00398719 +time/sampling batch (s) 0.279198 +time/saving (s) 0.00398836 +time/training (s) 6.36113 +time/epoch (s) 9.12082 +time/total (s) 6377.35 +Epoch -378 +---------------------------------- --------------- +2022-05-10 14:57:11.498526 PDT | [0] Epoch -377 finished +---------------------------------- --------------- +epoch -377 +replay_buffer/size 999033 +trainer/num train calls 624000 +trainer/Policy Loss -19.2548 +trainer/Log Pis Mean 23.8362 +trainer/Log Pis Std 12.957 +trainer/Log Pis Max 68.4618 +trainer/Log Pis Min -12.4167 +trainer/policy/mean Mean -0.0136139 +trainer/policy/mean Std 0.906211 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78913 +trainer/policy/normal/std Std 0.665333 +trainer/policy/normal/std Max 5.99467 +trainer/policy/normal/std Min 0.270288 +trainer/policy/normal/log_std Mean 0.984704 +trainer/policy/normal/log_std Std 0.323633 +trainer/policy/normal/log_std Max 1.79087 +trainer/policy/normal/log_std Min -1.30827 +eval/num steps total 620385 +eval/num paths total 624 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.327127 +eval/Actions Std 0.840632 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52037 +time/logging (s) 0.00378114 +time/sampling batch (s) 0.279347 +time/saving (s) 0.00356712 +time/training (s) 6.91736 +time/epoch (s) 9.72443 +time/total (s) 6387.08 +Epoch -377 +---------------------------------- --------------- +2022-05-10 14:57:21.713333 PDT | [0] Epoch -376 finished +---------------------------------- --------------- +epoch -376 +replay_buffer/size 999033 +trainer/num train calls 625000 +trainer/Policy Loss -19.0307 +trainer/Log Pis Mean 24.6905 +trainer/Log Pis Std 13.9555 +trainer/Log Pis Max 77.2434 +trainer/Log Pis Min -8.55398 +trainer/policy/mean Mean -0.0194016 +trainer/policy/mean Std 0.904544 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.86982 +trainer/policy/normal/std Std 0.672563 +trainer/policy/normal/std Max 5.97883 +trainer/policy/normal/std Min 0.301506 +trainer/policy/normal/log_std Mean 1.01449 +trainer/policy/normal/log_std Std 0.318178 +trainer/policy/normal/log_std Max 1.78823 +trainer/policy/normal/log_std Min -1.19896 +eval/num steps total 621385 +eval/num paths total 625 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0192922 +eval/Actions Std 0.929398 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64258 +time/logging (s) 0.0036919 +time/sampling batch (s) 0.279223 +time/saving (s) 0.00358354 +time/training (s) 7.26333 +time/epoch (s) 10.1924 +time/total (s) 6397.28 +Epoch -376 +---------------------------------- --------------- +2022-05-10 14:57:31.215630 PDT | [0] Epoch -375 finished +---------------------------------- --------------- +epoch -375 +replay_buffer/size 999033 +trainer/num train calls 626000 +trainer/Policy Loss -19.5468 +trainer/Log Pis Mean 24.2465 +trainer/Log Pis Std 12.9133 +trainer/Log Pis Max 60.6663 +trainer/Log Pis Min -8.90021 +trainer/policy/mean Mean -0.010297 +trainer/policy/mean Std 0.905274 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78721 +trainer/policy/normal/std Std 0.678605 +trainer/policy/normal/std Max 5.71702 +trainer/policy/normal/std Min 0.329001 +trainer/policy/normal/log_std Mean 0.980891 +trainer/policy/normal/log_std Std 0.338855 +trainer/policy/normal/log_std Max 1.74345 +trainer/policy/normal/log_std Min -1.1117 +eval/num steps total 622385 +eval/num paths total 626 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.2745 +eval/Actions Std 0.868535 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60968 +time/logging (s) 0.00367341 +time/sampling batch (s) 0.280124 +time/saving (s) 0.00346944 +time/training (s) 6.58325 +time/epoch (s) 9.4802 +time/total (s) 6406.76 +Epoch -375 +---------------------------------- --------------- +2022-05-10 14:57:41.297393 PDT | [0] Epoch -374 finished +---------------------------------- --------------- +epoch -374 +replay_buffer/size 999033 +trainer/num train calls 627000 +trainer/Policy Loss -20.3403 +trainer/Log Pis Mean 24.8626 +trainer/Log Pis Std 13.7252 +trainer/Log Pis Max 72.5319 +trainer/Log Pis Min -7.85287 +trainer/policy/mean Mean -0.0434867 +trainer/policy/mean Std 0.906095 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83251 +trainer/policy/normal/std Std 0.63794 +trainer/policy/normal/std Max 5.62827 +trainer/policy/normal/std Min 0.265597 +trainer/policy/normal/log_std Mean 1.00404 +trainer/policy/normal/log_std Std 0.309973 +trainer/policy/normal/log_std Max 1.7278 +trainer/policy/normal/log_std Min -1.32578 +eval/num steps total 623385 +eval/num paths total 627 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.409287 +eval/Actions Std 0.662481 +eval/Actions Max 0.999981 +eval/Actions Min -0.999964 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75866 +time/logging (s) 0.0044202 +time/sampling batch (s) 0.534219 +time/saving (s) 0.00382224 +time/training (s) 6.75896 +time/epoch (s) 10.0601 +time/total (s) 6416.82 +Epoch -374 +---------------------------------- --------------- +2022-05-10 14:57:50.926340 PDT | [0] Epoch -373 finished +---------------------------------- --------------- +epoch -373 +replay_buffer/size 999033 +trainer/num train calls 628000 +trainer/Policy Loss -19.3362 +trainer/Log Pis Mean 24.486 +trainer/Log Pis Std 13.2981 +trainer/Log Pis Max 71.0991 +trainer/Log Pis Min -7.52838 +trainer/policy/mean Mean -0.0227991 +trainer/policy/mean Std 0.906437 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84555 +trainer/policy/normal/std Std 0.6825 +trainer/policy/normal/std Max 6.77404 +trainer/policy/normal/std Min 0.244708 +trainer/policy/normal/log_std Mean 1.00241 +trainer/policy/normal/log_std Std 0.338027 +trainer/policy/normal/log_std Max 1.9131 +trainer/policy/normal/log_std Min -1.40769 +eval/num steps total 624385 +eval/num paths total 628 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.151893 +eval/Actions Std 0.877655 +eval/Actions Max 0.999988 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50959 +time/logging (s) 0.00425855 +time/sampling batch (s) 0.279895 +time/saving (s) 0.00407133 +time/training (s) 6.80797 +time/epoch (s) 9.60578 +time/total (s) 6426.43 +Epoch -373 +---------------------------------- --------------- +2022-05-10 14:58:00.918637 PDT | [0] Epoch -372 finished +---------------------------------- --------------- +epoch -372 +replay_buffer/size 999033 +trainer/num train calls 629000 +trainer/Policy Loss -20.4874 +trainer/Log Pis Mean 24.8762 +trainer/Log Pis Std 13.3096 +trainer/Log Pis Max 76.6538 +trainer/Log Pis Min -7.24124 +trainer/policy/mean Mean -0.0281639 +trainer/policy/mean Std 0.910474 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83541 +trainer/policy/normal/std Std 0.674872 +trainer/policy/normal/std Max 5.82818 +trainer/policy/normal/std Min 0.284464 +trainer/policy/normal/log_std Mean 1.00144 +trainer/policy/normal/log_std Std 0.323 +trainer/policy/normal/log_std Max 1.7627 +trainer/policy/normal/log_std Min -1.25715 +eval/num steps total 625385 +eval/num paths total 629 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.255585 +eval/Actions Std 0.890273 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60301 +time/logging (s) 0.00423614 +time/sampling batch (s) 0.280717 +time/saving (s) 0.00459551 +time/training (s) 7.07702 +time/epoch (s) 9.96957 +time/total (s) 6436.41 +Epoch -372 +---------------------------------- --------------- +2022-05-10 14:58:11.374584 PDT | [0] Epoch -371 finished +---------------------------------- --------------- +epoch -371 +replay_buffer/size 999033 +trainer/num train calls 630000 +trainer/Policy Loss -19.1943 +trainer/Log Pis Mean 24.7215 +trainer/Log Pis Std 13.4583 +trainer/Log Pis Max 74.0643 +trainer/Log Pis Min -10.8408 +trainer/policy/mean Mean -0.0393243 +trainer/policy/mean Std 0.905414 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.7842 +trainer/policy/normal/std Std 0.645103 +trainer/policy/normal/std Max 5.98345 +trainer/policy/normal/std Min 0.308755 +trainer/policy/normal/log_std Mean 0.984379 +trainer/policy/normal/log_std Std 0.319133 +trainer/policy/normal/log_std Max 1.789 +trainer/policy/normal/log_std Min -1.17521 +eval/num steps total 626385 +eval/num paths total 630 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0718438 +eval/Actions Std 0.88438 +eval/Actions Max 0.999999 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58528 +time/logging (s) 0.00418017 +time/sampling batch (s) 0.28189 +time/saving (s) 0.00420632 +time/training (s) 7.5575 +time/epoch (s) 10.4331 +time/total (s) 6446.84 +Epoch -371 +---------------------------------- --------------- +2022-05-10 14:58:21.364852 PDT | [0] Epoch -370 finished +---------------------------------- --------------- +epoch -370 +replay_buffer/size 999033 +trainer/num train calls 631000 +trainer/Policy Loss -20.5412 +trainer/Log Pis Mean 24.9379 +trainer/Log Pis Std 12.8596 +trainer/Log Pis Max 71.573 +trainer/Log Pis Min -2.40925 +trainer/policy/mean Mean -0.0481257 +trainer/policy/mean Std 0.907837 +trainer/policy/mean Max 1 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82493 +trainer/policy/normal/std Std 0.669536 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.256796 +trainer/policy/normal/log_std Mean 0.99805 +trainer/policy/normal/log_std Std 0.321581 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.35947 +eval/num steps total 627385 +eval/num paths total 631 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114804 +eval/Actions Std 0.914201 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59453 +time/logging (s) 0.00377744 +time/sampling batch (s) 0.284204 +time/saving (s) 0.00357237 +time/training (s) 7.08102 +time/epoch (s) 9.96711 +time/total (s) 6456.81 +Epoch -370 +---------------------------------- --------------- +2022-05-10 14:58:31.063294 PDT | [0] Epoch -369 finished +---------------------------------- --------------- +epoch -369 +replay_buffer/size 999033 +trainer/num train calls 632000 +trainer/Policy Loss -20.4944 +trainer/Log Pis Mean 24.3874 +trainer/Log Pis Std 13.2069 +trainer/Log Pis Max 65.1825 +trainer/Log Pis Min -9.23915 +trainer/policy/mean Mean -0.047609 +trainer/policy/mean Std 0.908251 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.77933 +trainer/policy/normal/std Std 0.644313 +trainer/policy/normal/std Max 5.50839 +trainer/policy/normal/std Min 0.255147 +trainer/policy/normal/log_std Mean 0.983375 +trainer/policy/normal/log_std Std 0.315472 +trainer/policy/normal/log_std Max 1.70627 +trainer/policy/normal/log_std Min -1.36591 +eval/num steps total 628385 +eval/num paths total 632 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0284856 +eval/Actions Std 0.903681 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60444 +time/logging (s) 0.00417976 +time/sampling batch (s) 0.278813 +time/saving (s) 0.00428016 +time/training (s) 6.78484 +time/epoch (s) 9.67656 +time/total (s) 6466.49 +Epoch -369 +---------------------------------- --------------- +2022-05-10 14:58:41.812473 PDT | [0] Epoch -368 finished +---------------------------------- --------------- +epoch -368 +replay_buffer/size 999033 +trainer/num train calls 633000 +trainer/Policy Loss -20.1662 +trainer/Log Pis Mean 24.9033 +trainer/Log Pis Std 13.2099 +trainer/Log Pis Max 67.7409 +trainer/Log Pis Min -7.82976 +trainer/policy/mean Mean -0.0385435 +trainer/policy/mean Std 0.903444 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82859 +trainer/policy/normal/std Std 0.653516 +trainer/policy/normal/std Max 5.81391 +trainer/policy/normal/std Min 0.314024 +trainer/policy/normal/log_std Mean 1.00159 +trainer/policy/normal/log_std Std 0.311231 +trainer/policy/normal/log_std Max 1.76025 +trainer/policy/normal/log_std Min -1.15829 +eval/num steps total 629385 +eval/num paths total 633 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0967146 +eval/Actions Std 0.909291 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6184 +time/logging (s) 0.00367965 +time/sampling batch (s) 0.528274 +time/saving (s) 0.00341896 +time/training (s) 7.57229 +time/epoch (s) 10.7261 +time/total (s) 6477.22 +Epoch -368 +---------------------------------- --------------- +2022-05-10 14:58:52.206454 PDT | [0] Epoch -367 finished +---------------------------------- --------------- +epoch -367 +replay_buffer/size 999033 +trainer/num train calls 634000 +trainer/Policy Loss -19.7403 +trainer/Log Pis Mean 23.3938 +trainer/Log Pis Std 13.2 +trainer/Log Pis Max 61.052 +trainer/Log Pis Min -10.1114 +trainer/policy/mean Mean -0.0239503 +trainer/policy/mean Std 0.907167 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82376 +trainer/policy/normal/std Std 0.670936 +trainer/policy/normal/std Max 5.77142 +trainer/policy/normal/std Min 0.312001 +trainer/policy/normal/log_std Mean 0.995255 +trainer/policy/normal/log_std Std 0.334972 +trainer/policy/normal/log_std Max 1.75292 +trainer/policy/normal/log_std Min -1.16475 +eval/num steps total 630385 +eval/num paths total 634 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.08853 +eval/Actions Std 0.902732 +eval/Actions Max 1 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57165 +time/logging (s) 0.00378188 +time/sampling batch (s) 0.280519 +time/saving (s) 0.00347064 +time/training (s) 7.51231 +time/epoch (s) 10.3717 +time/total (s) 6487.6 +Epoch -367 +---------------------------------- --------------- +2022-05-10 14:59:02.577513 PDT | [0] Epoch -366 finished +---------------------------------- --------------- +epoch -366 +replay_buffer/size 999033 +trainer/num train calls 635000 +trainer/Policy Loss -20.2099 +trainer/Log Pis Mean 24.9743 +trainer/Log Pis Std 13.2323 +trainer/Log Pis Max 69.2299 +trainer/Log Pis Min -8.81891 +trainer/policy/mean Mean -0.0277403 +trainer/policy/mean Std 0.901932 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.83179 +trainer/policy/normal/std Std 0.700063 +trainer/policy/normal/std Max 5.83788 +trainer/policy/normal/std Min 0.290012 +trainer/policy/normal/log_std Mean 0.995633 +trainer/policy/normal/log_std Std 0.343223 +trainer/policy/normal/log_std Max 1.76437 +trainer/policy/normal/log_std Min -1.23783 +eval/num steps total 631385 +eval/num paths total 635 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.276034 +eval/Actions Std 0.7793 +eval/Actions Max 0.999998 +eval/Actions Min -0.999966 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67314 +time/logging (s) 0.00391755 +time/sampling batch (s) 0.285521 +time/saving (s) 0.00408001 +time/training (s) 7.38004 +time/epoch (s) 10.3467 +time/total (s) 6497.95 +Epoch -366 +---------------------------------- --------------- +2022-05-10 14:59:12.734026 PDT | [0] Epoch -365 finished +---------------------------------- --------------- +epoch -365 +replay_buffer/size 999033 +trainer/num train calls 636000 +trainer/Policy Loss -19.3704 +trainer/Log Pis Mean 25.3516 +trainer/Log Pis Std 13.9166 +trainer/Log Pis Max 69.435 +trainer/Log Pis Min -9.11922 +trainer/policy/mean Mean -0.0394435 +trainer/policy/mean Std 0.909431 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80136 +trainer/policy/normal/std Std 0.640199 +trainer/policy/normal/std Max 5.50551 +trainer/policy/normal/std Min 0.274468 +trainer/policy/normal/log_std Mean 0.991734 +trainer/policy/normal/log_std Std 0.31526 +trainer/policy/normal/log_std Max 1.70575 +trainer/policy/normal/log_std Min -1.29292 +eval/num steps total 632385 +eval/num paths total 636 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0671207 +eval/Actions Std 0.859172 +eval/Actions Max 0.999999 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5065 +time/logging (s) 0.00366336 +time/sampling batch (s) 0.285266 +time/saving (s) 0.00341797 +time/training (s) 7.33467 +time/epoch (s) 10.1335 +time/total (s) 6508.08 +Epoch -365 +---------------------------------- --------------- +2022-05-10 14:59:22.820091 PDT | [0] Epoch -364 finished +---------------------------------- --------------- +epoch -364 +replay_buffer/size 999033 +trainer/num train calls 637000 +trainer/Policy Loss -20.5138 +trainer/Log Pis Mean 25.26 +trainer/Log Pis Std 12.4628 +trainer/Log Pis Max 69.9618 +trainer/Log Pis Min -5.25907 +trainer/policy/mean Mean -0.0432283 +trainer/policy/mean Std 0.911651 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.78944 +trainer/policy/normal/std Std 0.659726 +trainer/policy/normal/std Max 6.12604 +trainer/policy/normal/std Min 0.233406 +trainer/policy/normal/log_std Mean 0.985601 +trainer/policy/normal/log_std Std 0.320705 +trainer/policy/normal/log_std Max 1.81255 +trainer/policy/normal/log_std Min -1.45498 +eval/num steps total 633385 +eval/num paths total 637 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10464 +eval/Actions Std 0.880188 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6436 +time/logging (s) 0.00370517 +time/sampling batch (s) 0.286018 +time/saving (s) 0.00345589 +time/training (s) 7.12674 +time/epoch (s) 10.0635 +time/total (s) 6518.15 +Epoch -364 +---------------------------------- --------------- +2022-05-10 14:59:32.436224 PDT | [0] Epoch -363 finished +---------------------------------- --------------- +epoch -363 +replay_buffer/size 999033 +trainer/num train calls 638000 +trainer/Policy Loss -18.1286 +trainer/Log Pis Mean 23.3838 +trainer/Log Pis Std 13.0968 +trainer/Log Pis Max 64.4245 +trainer/Log Pis Min -7.71524 +trainer/policy/mean Mean -0.0364563 +trainer/policy/mean Std 0.90294 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78822 +trainer/policy/normal/std Std 0.661899 +trainer/policy/normal/std Max 5.46769 +trainer/policy/normal/std Min 0.318935 +trainer/policy/normal/log_std Mean 0.985419 +trainer/policy/normal/log_std Std 0.316896 +trainer/policy/normal/log_std Max 1.69886 +trainer/policy/normal/log_std Min -1.14277 +eval/num steps total 634385 +eval/num paths total 638 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0880681 +eval/Actions Std 0.910574 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58591 +time/logging (s) 0.00364591 +time/sampling batch (s) 0.283313 +time/saving (s) 0.00348992 +time/training (s) 6.71718 +time/epoch (s) 9.59354 +time/total (s) 6527.75 +Epoch -363 +---------------------------------- --------------- +2022-05-10 14:59:42.583012 PDT | [0] Epoch -362 finished +---------------------------------- --------------- +epoch -362 +replay_buffer/size 999033 +trainer/num train calls 639000 +trainer/Policy Loss -21.1532 +trainer/Log Pis Mean 25.2304 +trainer/Log Pis Std 13.0078 +trainer/Log Pis Max 63.3847 +trainer/Log Pis Min -6.57903 +trainer/policy/mean Mean -0.0370509 +trainer/policy/mean Std 0.905561 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.79676 +trainer/policy/normal/std Std 0.655432 +trainer/policy/normal/std Max 5.53925 +trainer/policy/normal/std Min 0.286479 +trainer/policy/normal/log_std Mean 0.988309 +trainer/policy/normal/log_std Std 0.321959 +trainer/policy/normal/log_std Max 1.71186 +trainer/policy/normal/log_std Min -1.25009 +eval/num steps total 635385 +eval/num paths total 639 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.223167 +eval/Actions Std 0.905656 +eval/Actions Max 0.999996 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59193 +time/logging (s) 0.00409383 +time/sampling batch (s) 0.532829 +time/saving (s) 0.00413106 +time/training (s) 6.99163 +time/epoch (s) 10.1246 +time/total (s) 6537.87 +Epoch -362 +---------------------------------- --------------- +2022-05-10 14:59:53.004296 PDT | [0] Epoch -361 finished +---------------------------------- --------------- +epoch -361 +replay_buffer/size 999033 +trainer/num train calls 640000 +trainer/Policy Loss -19.6717 +trainer/Log Pis Mean 24.4903 +trainer/Log Pis Std 13.9169 +trainer/Log Pis Max 66.4065 +trainer/Log Pis Min -7.56562 +trainer/policy/mean Mean -0.0382785 +trainer/policy/mean Std 0.91055 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85194 +trainer/policy/normal/std Std 0.652898 +trainer/policy/normal/std Max 6.23258 +trainer/policy/normal/std Min 0.281846 +trainer/policy/normal/log_std Mean 1.01052 +trainer/policy/normal/log_std Std 0.308596 +trainer/policy/normal/log_std Max 1.82979 +trainer/policy/normal/log_std Min -1.26639 +eval/num steps total 636385 +eval/num paths total 640 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.390228 +eval/Actions Std 0.859722 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60294 +time/logging (s) 0.00374598 +time/sampling batch (s) 0.28298 +time/saving (s) 0.00367452 +time/training (s) 7.50461 +time/epoch (s) 10.398 +time/total (s) 6548.27 +Epoch -361 +---------------------------------- --------------- +2022-05-10 15:00:02.975150 PDT | [0] Epoch -360 finished +---------------------------------- --------------- +epoch -360 +replay_buffer/size 999033 +trainer/num train calls 641000 +trainer/Policy Loss -19.9098 +trainer/Log Pis Mean 23.8546 +trainer/Log Pis Std 12.668 +trainer/Log Pis Max 75.4481 +trainer/Log Pis Min -6.86616 +trainer/policy/mean Mean -0.0273978 +trainer/policy/mean Std 0.906412 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.78449 +trainer/policy/normal/std Std 0.66576 +trainer/policy/normal/std Max 6.13378 +trainer/policy/normal/std Min 0.262248 +trainer/policy/normal/log_std Mean 0.981827 +trainer/policy/normal/log_std Std 0.330605 +trainer/policy/normal/log_std Max 1.81381 +trainer/policy/normal/log_std Min -1.33847 +eval/num steps total 637320 +eval/num paths total 641 +eval/path length Mean 935 +eval/path length Std 0 +eval/path length Max 935 +eval/path length Min 935 +eval/Rewards Mean 0.00106952 +eval/Rewards Std 0.032686 +eval/Rewards Max 1 +eval/Rewards Min 0 +eval/Returns Mean 1 +eval/Returns Std 0 +eval/Returns Max 1 +eval/Returns Min 1 +eval/Actions Mean 0.0368691 +eval/Actions Std 0.904758 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 1 +eval/normalized_score 100 +time/evaluation sampling (s) 2.57274 +time/logging (s) 0.00356693 +time/sampling batch (s) 0.283892 +time/saving (s) 0.0035155 +time/training (s) 7.08427 +time/epoch (s) 9.94799 +time/total (s) 6558.23 +Epoch -360 +---------------------------------- --------------- +2022-05-10 15:00:11.737842 PDT | [0] Epoch -359 finished +---------------------------------- --------------- +epoch -359 +replay_buffer/size 999033 +trainer/num train calls 642000 +trainer/Policy Loss -19.772 +trainer/Log Pis Mean 23.9657 +trainer/Log Pis Std 13.8252 +trainer/Log Pis Max 67.833 +trainer/Log Pis Min -8.88025 +trainer/policy/mean Mean -0.0323999 +trainer/policy/mean Std 0.903122 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81095 +trainer/policy/normal/std Std 0.666739 +trainer/policy/normal/std Max 6.39978 +trainer/policy/normal/std Min 0.337455 +trainer/policy/normal/log_std Mean 0.993223 +trainer/policy/normal/log_std Std 0.320713 +trainer/policy/normal/log_std Max 1.85626 +trainer/policy/normal/log_std Min -1.08632 +eval/num steps total 638320 +eval/num paths total 642 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.242752 +eval/Actions Std 0.91923 +eval/Actions Max 0.999993 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75687 +time/logging (s) 0.00372409 +time/sampling batch (s) 0.283076 +time/saving (s) 0.00356853 +time/training (s) 5.69315 +time/epoch (s) 8.74039 +time/total (s) 6566.97 +Epoch -359 +---------------------------------- --------------- +2022-05-10 15:00:21.598076 PDT | [0] Epoch -358 finished +---------------------------------- --------------- +epoch -358 +replay_buffer/size 999033 +trainer/num train calls 643000 +trainer/Policy Loss -20.0815 +trainer/Log Pis Mean 23.1911 +trainer/Log Pis Std 12.686 +trainer/Log Pis Max 78.7406 +trainer/Log Pis Min -9.62689 +trainer/policy/mean Mean -0.0248986 +trainer/policy/mean Std 0.907184 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.82719 +trainer/policy/normal/std Std 0.662387 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.275479 +trainer/policy/normal/log_std Mean 0.999844 +trainer/policy/normal/log_std Std 0.318263 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.28924 +eval/num steps total 639320 +eval/num paths total 643 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.35181 +eval/Actions Std 0.832385 +eval/Actions Max 0.999977 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66904 +time/logging (s) 0.00374155 +time/sampling batch (s) 0.533808 +time/saving (s) 0.00346835 +time/training (s) 6.62765 +time/epoch (s) 9.83771 +time/total (s) 6576.81 +Epoch -358 +---------------------------------- --------------- +2022-05-10 15:00:31.994557 PDT | [0] Epoch -357 finished +---------------------------------- --------------- +epoch -357 +replay_buffer/size 999033 +trainer/num train calls 644000 +trainer/Policy Loss -19.9579 +trainer/Log Pis Mean 24.6457 +trainer/Log Pis Std 12.5435 +trainer/Log Pis Max 71.9623 +trainer/Log Pis Min -9.21017 +trainer/policy/mean Mean -0.0393876 +trainer/policy/mean Std 0.909748 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.81528 +trainer/policy/normal/std Std 0.646951 +trainer/policy/normal/std Max 5.53482 +trainer/policy/normal/std Min 0.292168 +trainer/policy/normal/log_std Mean 0.995976 +trainer/policy/normal/log_std Std 0.318757 +trainer/policy/normal/log_std Max 1.71106 +trainer/policy/normal/log_std Min -1.23043 +eval/num steps total 640320 +eval/num paths total 644 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0523373 +eval/Actions Std 0.913651 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4618 +time/logging (s) 0.00366218 +time/sampling batch (s) 0.280859 +time/saving (s) 0.00338057 +time/training (s) 7.62442 +time/epoch (s) 10.3741 +time/total (s) 6587.19 +Epoch -357 +---------------------------------- --------------- +2022-05-10 15:00:42.378757 PDT | [0] Epoch -356 finished +---------------------------------- --------------- +epoch -356 +replay_buffer/size 999033 +trainer/num train calls 645000 +trainer/Policy Loss -19.6913 +trainer/Log Pis Mean 23.8891 +trainer/Log Pis Std 13.0954 +trainer/Log Pis Max 65.4858 +trainer/Log Pis Min -10.2077 +trainer/policy/mean Mean -0.024768 +trainer/policy/mean Std 0.901799 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80071 +trainer/policy/normal/std Std 0.673731 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.278032 +trainer/policy/normal/log_std Mean 0.987896 +trainer/policy/normal/log_std Std 0.327844 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.28002 +eval/num steps total 641320 +eval/num paths total 645 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141707 +eval/Actions Std 0.869082 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75717 +time/logging (s) 0.00409598 +time/sampling batch (s) 0.283027 +time/saving (s) 0.00413126 +time/training (s) 7.31376 +time/epoch (s) 10.3622 +time/total (s) 6597.55 +Epoch -356 +---------------------------------- --------------- +2022-05-10 15:00:53.686442 PDT | [0] Epoch -355 finished +---------------------------------- --------------- +epoch -355 +replay_buffer/size 999033 +trainer/num train calls 646000 +trainer/Policy Loss -19.5072 +trainer/Log Pis Mean 24.2783 +trainer/Log Pis Std 14.0035 +trainer/Log Pis Max 72.3055 +trainer/Log Pis Min -10.4951 +trainer/policy/mean Mean -0.030735 +trainer/policy/mean Std 0.905794 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82221 +trainer/policy/normal/std Std 0.667063 +trainer/policy/normal/std Max 5.86168 +trainer/policy/normal/std Min 0.278983 +trainer/policy/normal/log_std Mean 0.997355 +trainer/policy/normal/log_std Std 0.319636 +trainer/policy/normal/log_std Max 1.76844 +trainer/policy/normal/log_std Min -1.27661 +eval/num steps total 642320 +eval/num paths total 646 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0141426 +eval/Actions Std 0.909127 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61973 +time/logging (s) 0.00367881 +time/sampling batch (s) 0.53812 +time/saving (s) 0.00342268 +time/training (s) 8.11885 +time/epoch (s) 11.2838 +time/total (s) 6608.84 +Epoch -355 +---------------------------------- --------------- +2022-05-10 15:01:04.352679 PDT | [0] Epoch -354 finished +---------------------------------- --------------- +epoch -354 +replay_buffer/size 999033 +trainer/num train calls 647000 +trainer/Policy Loss -20.5988 +trainer/Log Pis Mean 24.921 +trainer/Log Pis Std 13.7977 +trainer/Log Pis Max 88.6998 +trainer/Log Pis Min -6.54793 +trainer/policy/mean Mean -0.0414544 +trainer/policy/mean Std 0.909304 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7957 +trainer/policy/normal/std Std 0.653003 +trainer/policy/normal/std Max 6.33547 +trainer/policy/normal/std Min 0.279447 +trainer/policy/normal/log_std Mean 0.990184 +trainer/policy/normal/log_std Std 0.30827 +trainer/policy/normal/log_std Max 1.84616 +trainer/policy/normal/log_std Min -1.27494 +eval/num steps total 643320 +eval/num paths total 647 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0758091 +eval/Actions Std 0.927384 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61132 +time/logging (s) 0.00377091 +time/sampling batch (s) 0.283278 +time/saving (s) 0.00350907 +time/training (s) 7.74201 +time/epoch (s) 10.6439 +time/total (s) 6619.49 +Epoch -354 +---------------------------------- --------------- +2022-05-10 15:01:14.966279 PDT | [0] Epoch -353 finished +---------------------------------- --------------- +epoch -353 +replay_buffer/size 999033 +trainer/num train calls 648000 +trainer/Policy Loss -19.9764 +trainer/Log Pis Mean 24.4195 +trainer/Log Pis Std 13.209 +trainer/Log Pis Max 73.6259 +trainer/Log Pis Min -6.00173 +trainer/policy/mean Mean -0.0436566 +trainer/policy/mean Std 0.907985 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.75343 +trainer/policy/normal/std Std 0.669982 +trainer/policy/normal/std Max 6.10188 +trainer/policy/normal/std Min 0.257128 +trainer/policy/normal/log_std Mean 0.968856 +trainer/policy/normal/log_std Std 0.337888 +trainer/policy/normal/log_std Max 1.8086 +trainer/policy/normal/log_std Min -1.35818 +eval/num steps total 644320 +eval/num paths total 648 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.123486 +eval/Actions Std 0.898665 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67032 +time/logging (s) 0.00368945 +time/sampling batch (s) 0.537808 +time/saving (s) 0.00345561 +time/training (s) 7.37536 +time/epoch (s) 10.5906 +time/total (s) 6630.08 +Epoch -353 +---------------------------------- --------------- +2022-05-10 15:01:25.421335 PDT | [0] Epoch -352 finished +---------------------------------- --------------- +epoch -352 +replay_buffer/size 999033 +trainer/num train calls 649000 +trainer/Policy Loss -19.5158 +trainer/Log Pis Mean 24.0051 +trainer/Log Pis Std 13.3603 +trainer/Log Pis Max 77.4972 +trainer/Log Pis Min -7.95192 +trainer/policy/mean Mean -0.0616407 +trainer/policy/mean Std 0.9056 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75994 +trainer/policy/normal/std Std 0.66794 +trainer/policy/normal/std Max 5.96946 +trainer/policy/normal/std Min 0.278402 +trainer/policy/normal/log_std Mean 0.972321 +trainer/policy/normal/log_std Std 0.332401 +trainer/policy/normal/log_std Max 1.78666 +trainer/policy/normal/log_std Min -1.27869 +eval/num steps total 645320 +eval/num paths total 649 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0216091 +eval/Actions Std 0.908268 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56925 +time/logging (s) 0.0036446 +time/sampling batch (s) 0.281541 +time/saving (s) 0.00340194 +time/training (s) 7.57468 +time/epoch (s) 10.4325 +time/total (s) 6640.51 +Epoch -352 +---------------------------------- --------------- +2022-05-10 15:01:35.667551 PDT | [0] Epoch -351 finished +---------------------------------- --------------- +epoch -351 +replay_buffer/size 999033 +trainer/num train calls 650000 +trainer/Policy Loss -20.134 +trainer/Log Pis Mean 24.6149 +trainer/Log Pis Std 13.2643 +trainer/Log Pis Max 66.5374 +trainer/Log Pis Min -5.92743 +trainer/policy/mean Mean -0.0499601 +trainer/policy/mean Std 0.902826 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.84177 +trainer/policy/normal/std Std 0.659355 +trainer/policy/normal/std Max 5.92407 +trainer/policy/normal/std Min 0.291674 +trainer/policy/normal/log_std Mean 1.00613 +trainer/policy/normal/log_std Std 0.31137 +trainer/policy/normal/log_std Max 1.77902 +trainer/policy/normal/log_std Min -1.23212 +eval/num steps total 646320 +eval/num paths total 650 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135633 +eval/Actions Std 0.899507 +eval/Actions Max 0.999948 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65638 +time/logging (s) 0.00412652 +time/sampling batch (s) 0.529633 +time/saving (s) 0.00417081 +time/training (s) 7.03018 +time/epoch (s) 10.2245 +time/total (s) 6650.74 +Epoch -351 +---------------------------------- --------------- +2022-05-10 15:01:45.781819 PDT | [0] Epoch -350 finished +---------------------------------- --------------- +epoch -350 +replay_buffer/size 999033 +trainer/num train calls 651000 +trainer/Policy Loss -18.9274 +trainer/Log Pis Mean 23.1098 +trainer/Log Pis Std 12.8631 +trainer/Log Pis Max 56.5647 +trainer/Log Pis Min -9.01835 +trainer/policy/mean Mean -0.0372946 +trainer/policy/mean Std 0.905967 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81215 +trainer/policy/normal/std Std 0.674741 +trainer/policy/normal/std Max 6.21937 +trainer/policy/normal/std Min 0.239188 +trainer/policy/normal/log_std Mean 0.991009 +trainer/policy/normal/log_std Std 0.334603 +trainer/policy/normal/log_std Max 1.82767 +trainer/policy/normal/log_std Min -1.4305 +eval/num steps total 647320 +eval/num paths total 651 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0394039 +eval/Actions Std 0.904918 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59631 +time/logging (s) 0.003752 +time/sampling batch (s) 0.277322 +time/saving (s) 0.00341419 +time/training (s) 7.21053 +time/epoch (s) 10.0913 +time/total (s) 6660.84 +Epoch -350 +---------------------------------- --------------- +2022-05-10 15:01:56.181820 PDT | [0] Epoch -349 finished +---------------------------------- --------------- +epoch -349 +replay_buffer/size 999033 +trainer/num train calls 652000 +trainer/Policy Loss -20.5069 +trainer/Log Pis Mean 23.9134 +trainer/Log Pis Std 12.9704 +trainer/Log Pis Max 65.8056 +trainer/Log Pis Min -7.53055 +trainer/policy/mean Mean -0.0486291 +trainer/policy/mean Std 0.906697 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.74868 +trainer/policy/normal/std Std 0.658779 +trainer/policy/normal/std Max 7.30652 +trainer/policy/normal/std Min 0.300125 +trainer/policy/normal/log_std Mean 0.969068 +trainer/policy/normal/log_std Std 0.329344 +trainer/policy/normal/log_std Max 1.98877 +trainer/policy/normal/log_std Min -1.20356 +eval/num steps total 648320 +eval/num paths total 652 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.228401 +eval/Actions Std 0.867767 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62575 +time/logging (s) 0.00375449 +time/sampling batch (s) 0.529847 +time/saving (s) 0.00341955 +time/training (s) 7.21502 +time/epoch (s) 10.3778 +time/total (s) 6671.22 +Epoch -349 +---------------------------------- --------------- +2022-05-10 15:02:07.161723 PDT | [0] Epoch -348 finished +---------------------------------- --------------- +epoch -348 +replay_buffer/size 999033 +trainer/num train calls 653000 +trainer/Policy Loss -20.2935 +trainer/Log Pis Mean 24.6381 +trainer/Log Pis Std 12.9638 +trainer/Log Pis Max 69.5607 +trainer/Log Pis Min -8.54827 +trainer/policy/mean Mean -0.0305956 +trainer/policy/mean Std 0.90642 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80928 +trainer/policy/normal/std Std 0.660844 +trainer/policy/normal/std Max 5.83883 +trainer/policy/normal/std Min 0.324414 +trainer/policy/normal/log_std Mean 0.993358 +trainer/policy/normal/log_std Std 0.317366 +trainer/policy/normal/log_std Max 1.76453 +trainer/policy/normal/log_std Min -1.12574 +eval/num steps total 649320 +eval/num paths total 653 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.303933 +eval/Actions Std 0.827464 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61301 +time/logging (s) 0.00373963 +time/sampling batch (s) 0.530726 +time/saving (s) 0.0034822 +time/training (s) 7.80668 +time/epoch (s) 10.9576 +time/total (s) 6682.18 +Epoch -348 +---------------------------------- --------------- +2022-05-10 15:02:16.963405 PDT | [0] Epoch -347 finished +---------------------------------- --------------- +epoch -347 +replay_buffer/size 999033 +trainer/num train calls 654000 +trainer/Policy Loss -19.3402 +trainer/Log Pis Mean 25.7797 +trainer/Log Pis Std 13.7971 +trainer/Log Pis Max 64.6462 +trainer/Log Pis Min -4.71299 +trainer/policy/mean Mean -0.0203724 +trainer/policy/mean Std 0.909918 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.78937 +trainer/policy/normal/std Std 0.661458 +trainer/policy/normal/std Max 5.76837 +trainer/policy/normal/std Min 0.302776 +trainer/policy/normal/log_std Mean 0.985372 +trainer/policy/normal/log_std Std 0.320469 +trainer/policy/normal/log_std Max 1.75239 +trainer/policy/normal/log_std Min -1.19476 +eval/num steps total 650320 +eval/num paths total 654 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.357022 +eval/Actions Std 0.864734 +eval/Actions Max 0.999994 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69588 +time/logging (s) 0.00377979 +time/sampling batch (s) 0.78554 +time/saving (s) 0.00349064 +time/training (s) 6.29044 +time/epoch (s) 9.77913 +time/total (s) 6691.96 +Epoch -347 +---------------------------------- --------------- +2022-05-10 15:02:27.484166 PDT | [0] Epoch -346 finished +---------------------------------- --------------- +epoch -346 +replay_buffer/size 999033 +trainer/num train calls 655000 +trainer/Policy Loss -19.5952 +trainer/Log Pis Mean 24.2972 +trainer/Log Pis Std 12.5892 +trainer/Log Pis Max 60.4294 +trainer/Log Pis Min -9.39233 +trainer/policy/mean Mean -0.0477384 +trainer/policy/mean Std 0.904811 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79595 +trainer/policy/normal/std Std 0.66018 +trainer/policy/normal/std Max 6.41014 +trainer/policy/normal/std Min 0.280669 +trainer/policy/normal/log_std Mean 0.987925 +trainer/policy/normal/log_std Std 0.320888 +trainer/policy/normal/log_std Max 1.85788 +trainer/policy/normal/log_std Min -1.27058 +eval/num steps total 651320 +eval/num paths total 655 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.129424 +eval/Actions Std 0.908172 +eval/Actions Max 0.999988 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52735 +time/logging (s) 0.00433548 +time/sampling batch (s) 0.294485 +time/saving (s) 0.00422709 +time/training (s) 7.66749 +time/epoch (s) 10.4979 +time/total (s) 6702.46 +Epoch -346 +---------------------------------- --------------- +2022-05-10 15:02:38.554067 PDT | [0] Epoch -345 finished +---------------------------------- --------------- +epoch -345 +replay_buffer/size 999033 +trainer/num train calls 656000 +trainer/Policy Loss -21.9121 +trainer/Log Pis Mean 25.3063 +trainer/Log Pis Std 13.5924 +trainer/Log Pis Max 68.2604 +trainer/Log Pis Min -10.4017 +trainer/policy/mean Mean -0.0386263 +trainer/policy/mean Std 0.906386 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.78157 +trainer/policy/normal/std Std 0.663959 +trainer/policy/normal/std Max 5.56397 +trainer/policy/normal/std Min 0.323132 +trainer/policy/normal/log_std Mean 0.979385 +trainer/policy/normal/log_std Std 0.338957 +trainer/policy/normal/log_std Max 1.71631 +trainer/policy/normal/log_std Min -1.12969 +eval/num steps total 652320 +eval/num paths total 656 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.199729 +eval/Actions Std 0.895451 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58725 +time/logging (s) 0.00383748 +time/sampling batch (s) 0.293074 +time/saving (s) 0.00367976 +time/training (s) 8.1579 +time/epoch (s) 11.0457 +time/total (s) 6713.51 +Epoch -345 +---------------------------------- --------------- +2022-05-10 15:02:49.419480 PDT | [0] Epoch -344 finished +---------------------------------- --------------- +epoch -344 +replay_buffer/size 999033 +trainer/num train calls 657000 +trainer/Policy Loss -19.4799 +trainer/Log Pis Mean 24.5169 +trainer/Log Pis Std 13.6775 +trainer/Log Pis Max 82.772 +trainer/Log Pis Min -14.0492 +trainer/policy/mean Mean -0.0527152 +trainer/policy/mean Std 0.906587 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.84688 +trainer/policy/normal/std Std 0.668194 +trainer/policy/normal/std Max 5.94788 +trainer/policy/normal/std Min 0.30042 +trainer/policy/normal/log_std Mean 1.00567 +trainer/policy/normal/log_std Std 0.323651 +trainer/policy/normal/log_std Max 1.78303 +trainer/policy/normal/log_std Min -1.20257 +eval/num steps total 653320 +eval/num paths total 657 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0879801 +eval/Actions Std 0.903661 +eval/Actions Max 0.999974 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53401 +time/logging (s) 0.00380368 +time/sampling batch (s) 0.29239 +time/saving (s) 0.00354567 +time/training (s) 8.00818 +time/epoch (s) 10.8419 +time/total (s) 6724.35 +Epoch -344 +---------------------------------- --------------- +2022-05-10 15:02:59.547867 PDT | [0] Epoch -343 finished +---------------------------------- --------------- +epoch -343 +replay_buffer/size 999033 +trainer/num train calls 658000 +trainer/Policy Loss -19.084 +trainer/Log Pis Mean 23.4963 +trainer/Log Pis Std 12.5241 +trainer/Log Pis Max 66.5426 +trainer/Log Pis Min -6.0761 +trainer/policy/mean Mean -0.0533553 +trainer/policy/mean Std 0.903802 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8372 +trainer/policy/normal/std Std 0.661152 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.295662 +trainer/policy/normal/log_std Mean 1.00499 +trainer/policy/normal/log_std Std 0.308472 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.21854 +eval/num steps total 654320 +eval/num paths total 658 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.072447 +eval/Actions Std 0.908083 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58521 +time/logging (s) 0.00390148 +time/sampling batch (s) 0.291791 +time/saving (s) 0.00353386 +time/training (s) 7.22077 +time/epoch (s) 10.1052 +time/total (s) 6734.46 +Epoch -343 +---------------------------------- --------------- +2022-05-10 15:03:10.413370 PDT | [0] Epoch -342 finished +---------------------------------- --------------- +epoch -342 +replay_buffer/size 999033 +trainer/num train calls 659000 +trainer/Policy Loss -19.2459 +trainer/Log Pis Mean 23.8958 +trainer/Log Pis Std 13.1282 +trainer/Log Pis Max 76.8181 +trainer/Log Pis Min -9.46652 +trainer/policy/mean Mean -0.0178968 +trainer/policy/mean Std 0.906075 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78276 +trainer/policy/normal/std Std 0.650904 +trainer/policy/normal/std Max 5.80039 +trainer/policy/normal/std Min 0.290146 +trainer/policy/normal/log_std Mean 0.984004 +trainer/policy/normal/log_std Std 0.316455 +trainer/policy/normal/log_std Max 1.75792 +trainer/policy/normal/log_std Min -1.23737 +eval/num steps total 655320 +eval/num paths total 659 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.160493 +eval/Actions Std 0.886388 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80988 +time/logging (s) 0.00389464 +time/sampling batch (s) 0.2886 +time/saving (s) 0.00348185 +time/training (s) 7.7364 +time/epoch (s) 10.8423 +time/total (s) 6745.31 +Epoch -342 +---------------------------------- --------------- +2022-05-10 15:03:20.458711 PDT | [0] Epoch -341 finished +---------------------------------- --------------- +epoch -341 +replay_buffer/size 999033 +trainer/num train calls 660000 +trainer/Policy Loss -19.6663 +trainer/Log Pis Mean 23.4791 +trainer/Log Pis Std 13.249 +trainer/Log Pis Max 81.249 +trainer/Log Pis Min -4.45182 +trainer/policy/mean Mean -0.0359721 +trainer/policy/mean Std 0.908707 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76314 +trainer/policy/normal/std Std 0.659523 +trainer/policy/normal/std Max 5.32809 +trainer/policy/normal/std Min 0.282963 +trainer/policy/normal/log_std Mean 0.973762 +trainer/policy/normal/log_std Std 0.333746 +trainer/policy/normal/log_std Max 1.67299 +trainer/policy/normal/log_std Min -1.26244 +eval/num steps total 656320 +eval/num paths total 660 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0864451 +eval/Actions Std 0.916036 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55638 +time/logging (s) 0.00405004 +time/sampling batch (s) 0.281325 +time/saving (s) 0.00410023 +time/training (s) 7.17704 +time/epoch (s) 10.0229 +time/total (s) 6755.33 +Epoch -341 +---------------------------------- --------------- +2022-05-10 15:03:31.423469 PDT | [0] Epoch -340 finished +---------------------------------- --------------- +epoch -340 +replay_buffer/size 999033 +trainer/num train calls 661000 +trainer/Policy Loss -20.2771 +trainer/Log Pis Mean 24.4542 +trainer/Log Pis Std 12.8909 +trainer/Log Pis Max 61.6522 +trainer/Log Pis Min -7.92565 +trainer/policy/mean Mean -0.0614502 +trainer/policy/mean Std 0.902369 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.75442 +trainer/policy/normal/std Std 0.640668 +trainer/policy/normal/std Max 5.24088 +trainer/policy/normal/std Min 0.259236 +trainer/policy/normal/log_std Mean 0.972386 +trainer/policy/normal/log_std Std 0.326075 +trainer/policy/normal/log_std Max 1.65649 +trainer/policy/normal/log_std Min -1.35002 +eval/num steps total 657320 +eval/num paths total 661 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0879043 +eval/Actions Std 0.911934 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65264 +time/logging (s) 0.00380385 +time/sampling batch (s) 0.541428 +time/saving (s) 0.00377219 +time/training (s) 7.73952 +time/epoch (s) 10.9412 +time/total (s) 6766.28 +Epoch -340 +---------------------------------- --------------- +2022-05-10 15:03:41.370778 PDT | [0] Epoch -339 finished +---------------------------------- ---------------- +epoch -339 +replay_buffer/size 999033 +trainer/num train calls 662000 +trainer/Policy Loss -19.8655 +trainer/Log Pis Mean 24.0554 +trainer/Log Pis Std 12.5829 +trainer/Log Pis Max 66.4793 +trainer/Log Pis Min -4.91272 +trainer/policy/mean Mean -0.017059 +trainer/policy/mean Std 0.904123 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84516 +trainer/policy/normal/std Std 0.658769 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.267539 +trainer/policy/normal/log_std Mean 1.00664 +trainer/policy/normal/log_std Std 0.317406 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.31849 +eval/num steps total 658320 +eval/num paths total 662 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.000841333 +eval/Actions Std 0.911406 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54817 +time/logging (s) 0.00368036 +time/sampling batch (s) 0.532256 +time/saving (s) 0.0033742 +time/training (s) 6.8369 +time/epoch (s) 9.92439 +time/total (s) 6776.2 +Epoch -339 +---------------------------------- ---------------- +2022-05-10 15:03:52.086950 PDT | [0] Epoch -338 finished +---------------------------------- --------------- +epoch -338 +replay_buffer/size 999033 +trainer/num train calls 663000 +trainer/Policy Loss -18.9905 +trainer/Log Pis Mean 24.1533 +trainer/Log Pis Std 13.209 +trainer/Log Pis Max 63.1508 +trainer/Log Pis Min -6.69427 +trainer/policy/mean Mean -0.0439347 +trainer/policy/mean Std 0.906491 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.7885 +trainer/policy/normal/std Std 0.637817 +trainer/policy/normal/std Max 6.65347 +trainer/policy/normal/std Min 0.287668 +trainer/policy/normal/log_std Mean 0.988232 +trainer/policy/normal/log_std Std 0.307082 +trainer/policy/normal/log_std Max 1.89514 +trainer/policy/normal/log_std Min -1.24595 +eval/num steps total 659320 +eval/num paths total 663 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.111016 +eval/Actions Std 0.901005 +eval/Actions Max 0.999995 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53752 +time/logging (s) 0.00379577 +time/sampling batch (s) 0.539022 +time/saving (s) 0.00342118 +time/training (s) 7.60998 +time/epoch (s) 10.6937 +time/total (s) 6786.9 +Epoch -338 +---------------------------------- --------------- +2022-05-10 15:04:02.870014 PDT | [0] Epoch -337 finished +---------------------------------- --------------- +epoch -337 +replay_buffer/size 999033 +trainer/num train calls 664000 +trainer/Policy Loss -19.3027 +trainer/Log Pis Mean 24.8034 +trainer/Log Pis Std 13.5654 +trainer/Log Pis Max 76.2719 +trainer/Log Pis Min -8.80659 +trainer/policy/mean Mean -0.0150605 +trainer/policy/mean Std 0.90354 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80795 +trainer/policy/normal/std Std 0.703657 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.29117 +trainer/policy/normal/log_std Mean 0.985618 +trainer/policy/normal/log_std Std 0.349905 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.23385 +eval/num steps total 660320 +eval/num paths total 664 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.136253 +eval/Actions Std 0.910378 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65342 +time/logging (s) 0.00385436 +time/sampling batch (s) 0.536146 +time/saving (s) 0.00426887 +time/training (s) 7.56246 +time/epoch (s) 10.7601 +time/total (s) 6797.66 +Epoch -337 +---------------------------------- --------------- +2022-05-10 15:04:12.752350 PDT | [0] Epoch -336 finished +---------------------------------- --------------- +epoch -336 +replay_buffer/size 999033 +trainer/num train calls 665000 +trainer/Policy Loss -19.1874 +trainer/Log Pis Mean 24.2175 +trainer/Log Pis Std 13.2378 +trainer/Log Pis Max 68.9325 +trainer/Log Pis Min -7.03999 +trainer/policy/mean Mean -0.0470346 +trainer/policy/mean Std 0.903819 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81392 +trainer/policy/normal/std Std 0.638658 +trainer/policy/normal/std Max 6.40397 +trainer/policy/normal/std Min 0.232306 +trainer/policy/normal/log_std Mean 0.997187 +trainer/policy/normal/log_std Std 0.31114 +trainer/policy/normal/log_std Max 1.85692 +trainer/policy/normal/log_std Min -1.4597 +eval/num steps total 661320 +eval/num paths total 665 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0259861 +eval/Actions Std 0.950299 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45071 +time/logging (s) 0.00415169 +time/sampling batch (s) 0.531099 +time/saving (s) 0.00414825 +time/training (s) 6.86929 +time/epoch (s) 9.8594 +time/total (s) 6807.53 +Epoch -336 +---------------------------------- --------------- +2022-05-10 15:04:23.026756 PDT | [0] Epoch -335 finished +---------------------------------- --------------- +epoch -335 +replay_buffer/size 999033 +trainer/num train calls 666000 +trainer/Policy Loss -19.742 +trainer/Log Pis Mean 25.45 +trainer/Log Pis Std 13.7864 +trainer/Log Pis Max 72.2008 +trainer/Log Pis Min -10.6865 +trainer/policy/mean Mean -0.0429061 +trainer/policy/mean Std 0.911489 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79181 +trainer/policy/normal/std Std 0.655301 +trainer/policy/normal/std Max 5.26889 +trainer/policy/normal/std Min 0.30365 +trainer/policy/normal/log_std Mean 0.984949 +trainer/policy/normal/log_std Std 0.330114 +trainer/policy/normal/log_std Max 1.66182 +trainer/policy/normal/log_std Min -1.19188 +eval/num steps total 662320 +eval/num paths total 666 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0249793 +eval/Actions Std 0.910783 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5967 +time/logging (s) 0.00373219 +time/sampling batch (s) 0.285845 +time/saving (s) 0.0037592 +time/training (s) 7.3606 +time/epoch (s) 10.2506 +time/total (s) 6817.78 +Epoch -335 +---------------------------------- --------------- +2022-05-10 15:04:33.481426 PDT | [0] Epoch -334 finished +---------------------------------- --------------- +epoch -334 +replay_buffer/size 999033 +trainer/num train calls 667000 +trainer/Policy Loss -19.5818 +trainer/Log Pis Mean 24.1667 +trainer/Log Pis Std 13.7401 +trainer/Log Pis Max 66.9175 +trainer/Log Pis Min -5.01925 +trainer/policy/mean Mean -0.0484388 +trainer/policy/mean Std 0.904713 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.75258 +trainer/policy/normal/std Std 0.691541 +trainer/policy/normal/std Max 5.42438 +trainer/policy/normal/std Min 0.226704 +trainer/policy/normal/log_std Mean 0.963237 +trainer/policy/normal/log_std Std 0.363446 +trainer/policy/normal/log_std Max 1.6909 +trainer/policy/normal/log_std Min -1.48411 +eval/num steps total 663320 +eval/num paths total 667 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.217903 +eval/Actions Std 0.946634 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71908 +time/logging (s) 0.00372289 +time/sampling batch (s) 0.533972 +time/saving (s) 0.00348215 +time/training (s) 7.17133 +time/epoch (s) 10.4316 +time/total (s) 6828.21 +Epoch -334 +---------------------------------- --------------- +2022-05-10 15:04:43.403896 PDT | [0] Epoch -333 finished +---------------------------------- --------------- +epoch -333 +replay_buffer/size 999033 +trainer/num train calls 668000 +trainer/Policy Loss -19.5277 +trainer/Log Pis Mean 23.9025 +trainer/Log Pis Std 12.5134 +trainer/Log Pis Max 59.482 +trainer/Log Pis Min -6.30216 +trainer/policy/mean Mean -0.0352952 +trainer/policy/mean Std 0.903728 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76566 +trainer/policy/normal/std Std 0.706425 +trainer/policy/normal/std Max 7.35389 +trainer/policy/normal/std Min 0.229503 +trainer/policy/normal/log_std Mean 0.965549 +trainer/policy/normal/log_std Std 0.373619 +trainer/policy/normal/log_std Max 1.99523 +trainer/policy/normal/log_std Min -1.47184 +eval/num steps total 664320 +eval/num paths total 668 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0955518 +eval/Actions Std 0.892397 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68862 +time/logging (s) 0.00370644 +time/sampling batch (s) 0.529762 +time/saving (s) 0.00343146 +time/training (s) 6.67455 +time/epoch (s) 9.90007 +time/total (s) 6838.12 +Epoch -333 +---------------------------------- --------------- +2022-05-10 15:04:53.874480 PDT | [0] Epoch -332 finished +---------------------------------- --------------- +epoch -332 +replay_buffer/size 999033 +trainer/num train calls 669000 +trainer/Policy Loss -18.9893 +trainer/Log Pis Mean 25.3468 +trainer/Log Pis Std 13.8827 +trainer/Log Pis Max 84.1341 +trainer/Log Pis Min -3.98699 +trainer/policy/mean Mean -0.0597252 +trainer/policy/mean Std 0.906971 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.90213 +trainer/policy/normal/std Std 0.656217 +trainer/policy/normal/std Max 5.69533 +trainer/policy/normal/std Min 0.284002 +trainer/policy/normal/log_std Mean 1.02921 +trainer/policy/normal/log_std Std 0.30253 +trainer/policy/normal/log_std Max 1.73965 +trainer/policy/normal/log_std Min -1.25877 +eval/num steps total 665320 +eval/num paths total 669 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.510682 +eval/Actions Std 0.773707 +eval/Actions Max 0.999994 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76809 +time/logging (s) 0.00377158 +time/sampling batch (s) 0.533806 +time/saving (s) 0.00345979 +time/training (s) 7.13883 +time/epoch (s) 10.448 +time/total (s) 6848.57 +Epoch -332 +---------------------------------- --------------- +2022-05-10 15:05:03.668840 PDT | [0] Epoch -331 finished +---------------------------------- --------------- +epoch -331 +replay_buffer/size 999033 +trainer/num train calls 670000 +trainer/Policy Loss -19.5678 +trainer/Log Pis Mean 25.3628 +trainer/Log Pis Std 13.8173 +trainer/Log Pis Max 74.5374 +trainer/Log Pis Min -4.36107 +trainer/policy/mean Mean -0.0512645 +trainer/policy/mean Std 0.90405 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80105 +trainer/policy/normal/std Std 0.668048 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.292601 +trainer/policy/normal/log_std Mean 0.989541 +trainer/policy/normal/log_std Std 0.320243 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22895 +eval/num steps total 666320 +eval/num paths total 670 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0994874 +eval/Actions Std 0.905017 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5999 +time/logging (s) 0.00385683 +time/sampling batch (s) 0.279745 +time/saving (s) 0.00349576 +time/training (s) 6.88503 +time/epoch (s) 9.77202 +time/total (s) 6858.34 +Epoch -331 +---------------------------------- --------------- +2022-05-10 15:05:13.462134 PDT | [0] Epoch -330 finished +---------------------------------- --------------- +epoch -330 +replay_buffer/size 999033 +trainer/num train calls 671000 +trainer/Policy Loss -19.3794 +trainer/Log Pis Mean 22.9284 +trainer/Log Pis Std 12.409 +trainer/Log Pis Max 76.3427 +trainer/Log Pis Min -13.5788 +trainer/policy/mean Mean -0.039302 +trainer/policy/mean Std 0.905548 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75336 +trainer/policy/normal/std Std 0.647671 +trainer/policy/normal/std Max 5.76692 +trainer/policy/normal/std Min 0.312869 +trainer/policy/normal/log_std Mean 0.972202 +trainer/policy/normal/log_std Std 0.324024 +trainer/policy/normal/log_std Max 1.75214 +trainer/policy/normal/log_std Min -1.16197 +eval/num steps total 667320 +eval/num paths total 671 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.037975 +eval/Actions Std 0.90916 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5118 +time/logging (s) 0.00406788 +time/sampling batch (s) 0.280114 +time/saving (s) 0.00417974 +time/training (s) 6.97051 +time/epoch (s) 9.77067 +time/total (s) 6868.12 +Epoch -330 +---------------------------------- --------------- +2022-05-10 15:05:23.414549 PDT | [0] Epoch -329 finished +---------------------------------- --------------- +epoch -329 +replay_buffer/size 999033 +trainer/num train calls 672000 +trainer/Policy Loss -19.7099 +trainer/Log Pis Mean 24.7663 +trainer/Log Pis Std 12.6366 +trainer/Log Pis Max 60.7855 +trainer/Log Pis Min -3.93846 +trainer/policy/mean Mean -0.0366539 +trainer/policy/mean Std 0.910488 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.79813 +trainer/policy/normal/std Std 0.639414 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.266982 +trainer/policy/normal/log_std Mean 0.990836 +trainer/policy/normal/log_std Std 0.314217 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.32058 +eval/num steps total 668320 +eval/num paths total 672 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.21664 +eval/Actions Std 0.922807 +eval/Actions Max 0.999995 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6044 +time/logging (s) 0.00381587 +time/sampling batch (s) 0.279181 +time/saving (s) 0.00357141 +time/training (s) 7.03844 +time/epoch (s) 9.92941 +time/total (s) 6878.05 +Epoch -329 +---------------------------------- --------------- +2022-05-10 15:05:33.033812 PDT | [0] Epoch -328 finished +---------------------------------- --------------- +epoch -328 +replay_buffer/size 999033 +trainer/num train calls 673000 +trainer/Policy Loss -19.0507 +trainer/Log Pis Mean 23.5804 +trainer/Log Pis Std 13.2789 +trainer/Log Pis Max 66.3531 +trainer/Log Pis Min -5.72056 +trainer/policy/mean Mean -0.0319825 +trainer/policy/mean Std 0.907216 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79364 +trainer/policy/normal/std Std 0.678768 +trainer/policy/normal/std Max 6.88474 +trainer/policy/normal/std Min 0.260189 +trainer/policy/normal/log_std Mean 0.983415 +trainer/policy/normal/log_std Std 0.338056 +trainer/policy/normal/log_std Max 1.92931 +trainer/policy/normal/log_std Min -1.34635 +eval/num steps total 669320 +eval/num paths total 673 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101114 +eval/Actions Std 0.919443 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67803 +time/logging (s) 0.00372893 +time/sampling batch (s) 0.278783 +time/saving (s) 0.00340821 +time/training (s) 6.63278 +time/epoch (s) 9.59673 +time/total (s) 6887.65 +Epoch -328 +---------------------------------- --------------- +2022-05-10 15:05:42.893102 PDT | [0] Epoch -327 finished +---------------------------------- --------------- +epoch -327 +replay_buffer/size 999033 +trainer/num train calls 674000 +trainer/Policy Loss -20.1354 +trainer/Log Pis Mean 25.2482 +trainer/Log Pis Std 12.927 +trainer/Log Pis Max 63.2854 +trainer/Log Pis Min -8.03702 +trainer/policy/mean Mean -0.0244153 +trainer/policy/mean Std 0.906638 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81571 +trainer/policy/normal/std Std 0.657886 +trainer/policy/normal/std Max 6.54906 +trainer/policy/normal/std Min 0.281858 +trainer/policy/normal/log_std Mean 0.995281 +trainer/policy/normal/log_std Std 0.321077 +trainer/policy/normal/log_std Max 1.87932 +trainer/policy/normal/log_std Min -1.26635 +eval/num steps total 670320 +eval/num paths total 674 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0933355 +eval/Actions Std 0.914417 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69807 +time/logging (s) 0.00381812 +time/sampling batch (s) 0.279126 +time/saving (s) 0.00342675 +time/training (s) 6.85258 +time/epoch (s) 9.83702 +time/total (s) 6897.49 +Epoch -327 +---------------------------------- --------------- +2022-05-10 15:05:53.902821 PDT | [0] Epoch -326 finished +---------------------------------- --------------- +epoch -326 +replay_buffer/size 999033 +trainer/num train calls 675000 +trainer/Policy Loss -19.6458 +trainer/Log Pis Mean 23.97 +trainer/Log Pis Std 13.2401 +trainer/Log Pis Max 63.0586 +trainer/Log Pis Min -7.55044 +trainer/policy/mean Mean -0.0372147 +trainer/policy/mean Std 0.903212 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75181 +trainer/policy/normal/std Std 0.678071 +trainer/policy/normal/std Max 6.46835 +trainer/policy/normal/std Min 0.261781 +trainer/policy/normal/log_std Mean 0.968521 +trainer/policy/normal/log_std Std 0.334874 +trainer/policy/normal/log_std Max 1.86692 +trainer/policy/normal/log_std Min -1.34025 +eval/num steps total 671320 +eval/num paths total 675 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.199493 +eval/Actions Std 0.781943 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66 +time/logging (s) 0.00382266 +time/sampling batch (s) 0.280668 +time/saving (s) 0.00342717 +time/training (s) 8.03928 +time/epoch (s) 10.9872 +time/total (s) 6908.48 +Epoch -326 +---------------------------------- --------------- +2022-05-10 15:06:04.022264 PDT | [0] Epoch -325 finished +---------------------------------- --------------- +epoch -325 +replay_buffer/size 999033 +trainer/num train calls 676000 +trainer/Policy Loss -21.0669 +trainer/Log Pis Mean 25.099 +trainer/Log Pis Std 13.4842 +trainer/Log Pis Max 69.8408 +trainer/Log Pis Min -10.1689 +trainer/policy/mean Mean -0.0263015 +trainer/policy/mean Std 0.907072 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81724 +trainer/policy/normal/std Std 0.667685 +trainer/policy/normal/std Max 6.29906 +trainer/policy/normal/std Min 0.281741 +trainer/policy/normal/log_std Mean 0.994245 +trainer/policy/normal/log_std Std 0.327785 +trainer/policy/normal/log_std Max 1.8404 +trainer/policy/normal/log_std Min -1.26677 +eval/num steps total 672320 +eval/num paths total 676 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.417888 +eval/Actions Std 0.85062 +eval/Actions Max 0.999979 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57776 +time/logging (s) 0.00423937 +time/sampling batch (s) 0.278717 +time/saving (s) 0.00412062 +time/training (s) 7.23246 +time/epoch (s) 10.0973 +time/total (s) 6918.58 +Epoch -325 +---------------------------------- --------------- +2022-05-10 15:06:13.867732 PDT | [0] Epoch -324 finished +---------------------------------- --------------- +epoch -324 +replay_buffer/size 999033 +trainer/num train calls 677000 +trainer/Policy Loss -19.4802 +trainer/Log Pis Mean 24.6619 +trainer/Log Pis Std 13.7302 +trainer/Log Pis Max 65.3481 +trainer/Log Pis Min -8.81891 +trainer/policy/mean Mean -0.0321442 +trainer/policy/mean Std 0.906127 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79552 +trainer/policy/normal/std Std 0.661013 +trainer/policy/normal/std Max 5.81421 +trainer/policy/normal/std Min 0.347277 +trainer/policy/normal/log_std Mean 0.988583 +trainer/policy/normal/log_std Std 0.315083 +trainer/policy/normal/log_std Max 1.76031 +trainer/policy/normal/log_std Min -1.05763 +eval/num steps total 673320 +eval/num paths total 677 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00533261 +eval/Actions Std 0.946169 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63789 +time/logging (s) 0.00373008 +time/sampling batch (s) 0.2854 +time/saving (s) 0.00355185 +time/training (s) 6.89128 +time/epoch (s) 9.82185 +time/total (s) 6928.4 +Epoch -324 +---------------------------------- --------------- +2022-05-10 15:06:23.658602 PDT | [0] Epoch -323 finished +---------------------------------- --------------- +epoch -323 +replay_buffer/size 999033 +trainer/num train calls 678000 +trainer/Policy Loss -19.9005 +trainer/Log Pis Mean 25.0664 +trainer/Log Pis Std 13.3932 +trainer/Log Pis Max 90.993 +trainer/Log Pis Min -4.04378 +trainer/policy/mean Mean -0.0510348 +trainer/policy/mean Std 0.90703 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.82202 +trainer/policy/normal/std Std 0.632939 +trainer/policy/normal/std Max 5.15856 +trainer/policy/normal/std Min 0.245621 +trainer/policy/normal/log_std Mean 1.00126 +trainer/policy/normal/log_std Std 0.303734 +trainer/policy/normal/log_std Max 1.64066 +trainer/policy/normal/log_std Min -1.40396 +eval/num steps total 674320 +eval/num paths total 678 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.231233 +eval/Actions Std 0.943798 +eval/Actions Max 0.99999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64539 +time/logging (s) 0.0037896 +time/sampling batch (s) 0.282714 +time/saving (s) 0.00396033 +time/training (s) 6.83239 +time/epoch (s) 9.76824 +time/total (s) 6938.17 +Epoch -323 +---------------------------------- --------------- +2022-05-10 15:06:33.553304 PDT | [0] Epoch -322 finished +---------------------------------- --------------- +epoch -322 +replay_buffer/size 999033 +trainer/num train calls 679000 +trainer/Policy Loss -18.44 +trainer/Log Pis Mean 23.9314 +trainer/Log Pis Std 13.5055 +trainer/Log Pis Max 86.9868 +trainer/Log Pis Min -6.88504 +trainer/policy/mean Mean -0.0373957 +trainer/policy/mean Std 0.906288 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77697 +trainer/policy/normal/std Std 0.663224 +trainer/policy/normal/std Max 5.55566 +trainer/policy/normal/std Min 0.274153 +trainer/policy/normal/log_std Mean 0.978974 +trainer/policy/normal/log_std Std 0.331168 +trainer/policy/normal/log_std Max 1.71482 +trainer/policy/normal/log_std Min -1.29407 +eval/num steps total 675320 +eval/num paths total 679 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.126948 +eval/Actions Std 0.90638 +eval/Actions Max 0.999984 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.83314 +time/logging (s) 0.00363943 +time/sampling batch (s) 0.281383 +time/saving (s) 0.00343747 +time/training (s) 6.74914 +time/epoch (s) 9.87074 +time/total (s) 6948.05 +Epoch -322 +---------------------------------- --------------- +2022-05-10 15:06:43.639544 PDT | [0] Epoch -321 finished +---------------------------------- --------------- +epoch -321 +replay_buffer/size 999033 +trainer/num train calls 680000 +trainer/Policy Loss -18.3068 +trainer/Log Pis Mean 24.3178 +trainer/Log Pis Std 13.5151 +trainer/Log Pis Max 67.4855 +trainer/Log Pis Min -4.98475 +trainer/policy/mean Mean -0.0260997 +trainer/policy/mean Std 0.910133 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82528 +trainer/policy/normal/std Std 0.659657 +trainer/policy/normal/std Max 5.44658 +trainer/policy/normal/std Min 0.304067 +trainer/policy/normal/log_std Mean 0.99914 +trainer/policy/normal/log_std Std 0.31779 +trainer/policy/normal/log_std Max 1.69499 +trainer/policy/normal/log_std Min -1.19051 +eval/num steps total 676320 +eval/num paths total 680 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114459 +eval/Actions Std 0.935087 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53978 +time/logging (s) 0.00366484 +time/sampling batch (s) 0.281319 +time/saving (s) 0.00341289 +time/training (s) 7.23532 +time/epoch (s) 10.0635 +time/total (s) 6958.12 +Epoch -321 +---------------------------------- --------------- +2022-05-10 15:06:54.720421 PDT | [0] Epoch -320 finished +---------------------------------- --------------- +epoch -320 +replay_buffer/size 999033 +trainer/num train calls 681000 +trainer/Policy Loss -18.1654 +trainer/Log Pis Mean 25.1283 +trainer/Log Pis Std 14.5325 +trainer/Log Pis Max 74.3875 +trainer/Log Pis Min -7.5662 +trainer/policy/mean Mean -0.0319096 +trainer/policy/mean Std 0.906627 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.78656 +trainer/policy/normal/std Std 0.658749 +trainer/policy/normal/std Max 5.79656 +trainer/policy/normal/std Min 0.328598 +trainer/policy/normal/log_std Mean 0.983941 +trainer/policy/normal/log_std Std 0.323956 +trainer/policy/normal/log_std Max 1.75727 +trainer/policy/normal/log_std Min -1.11292 +eval/num steps total 677320 +eval/num paths total 681 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.207588 +eval/Actions Std 0.924412 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58031 +time/logging (s) 0.0043243 +time/sampling batch (s) 0.540569 +time/saving (s) 0.00427284 +time/training (s) 7.92935 +time/epoch (s) 11.0588 +time/total (s) 6969.18 +Epoch -320 +---------------------------------- --------------- +2022-05-10 15:07:04.397805 PDT | [0] Epoch -319 finished +---------------------------------- --------------- +epoch -319 +replay_buffer/size 999033 +trainer/num train calls 682000 +trainer/Policy Loss -18.8039 +trainer/Log Pis Mean 23.1282 +trainer/Log Pis Std 12.2938 +trainer/Log Pis Max 62.0733 +trainer/Log Pis Min -6.13925 +trainer/policy/mean Mean -0.038523 +trainer/policy/mean Std 0.906416 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.73567 +trainer/policy/normal/std Std 0.63406 +trainer/policy/normal/std Max 5.84265 +trainer/policy/normal/std Min 0.30045 +trainer/policy/normal/log_std Mean 0.967235 +trainer/policy/normal/log_std Std 0.316069 +trainer/policy/normal/log_std Max 1.76519 +trainer/policy/normal/log_std Min -1.20247 +eval/num steps total 678320 +eval/num paths total 682 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0224943 +eval/Actions Std 0.896363 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60346 +time/logging (s) 0.00395201 +time/sampling batch (s) 0.282085 +time/saving (s) 0.00414088 +time/training (s) 6.76023 +time/epoch (s) 9.65388 +time/total (s) 6978.83 +Epoch -319 +---------------------------------- --------------- +2022-05-10 15:07:14.247128 PDT | [0] Epoch -318 finished +---------------------------------- --------------- +epoch -318 +replay_buffer/size 999033 +trainer/num train calls 683000 +trainer/Policy Loss -19.761 +trainer/Log Pis Mean 23.9091 +trainer/Log Pis Std 13.4167 +trainer/Log Pis Max 67.0471 +trainer/Log Pis Min -5.54408 +trainer/policy/mean Mean -0.0529809 +trainer/policy/mean Std 0.905637 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.73958 +trainer/policy/normal/std Std 0.631076 +trainer/policy/normal/std Max 6.02962 +trainer/policy/normal/std Min 0.266911 +trainer/policy/normal/log_std Mean 0.968554 +trainer/policy/normal/log_std Std 0.318244 +trainer/policy/normal/log_std Max 1.79668 +trainer/policy/normal/log_std Min -1.32084 +eval/num steps total 679320 +eval/num paths total 683 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0918074 +eval/Actions Std 0.901503 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63148 +time/logging (s) 0.00394022 +time/sampling batch (s) 0.281315 +time/saving (s) 0.0041379 +time/training (s) 6.9053 +time/epoch (s) 9.82617 +time/total (s) 6988.66 +Epoch -318 +---------------------------------- --------------- +2022-05-10 15:07:24.090980 PDT | [0] Epoch -317 finished +---------------------------------- --------------- +epoch -317 +replay_buffer/size 999033 +trainer/num train calls 684000 +trainer/Policy Loss -19.9357 +trainer/Log Pis Mean 24.0514 +trainer/Log Pis Std 12.8569 +trainer/Log Pis Max 68.7924 +trainer/Log Pis Min -4.1281 +trainer/policy/mean Mean -0.0356031 +trainer/policy/mean Std 0.906656 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.73338 +trainer/policy/normal/std Std 0.638074 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.246687 +trainer/policy/normal/log_std Mean 0.966085 +trainer/policy/normal/log_std Std 0.317826 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.39963 +eval/num steps total 680320 +eval/num paths total 684 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.151752 +eval/Actions Std 0.864032 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6758 +time/logging (s) 0.00418871 +time/sampling batch (s) 0.281921 +time/saving (s) 0.00438253 +time/training (s) 6.85459 +time/epoch (s) 9.82088 +time/total (s) 6998.49 +Epoch -317 +---------------------------------- --------------- +2022-05-10 15:07:34.824531 PDT | [0] Epoch -316 finished +---------------------------------- --------------- +epoch -316 +replay_buffer/size 999033 +trainer/num train calls 685000 +trainer/Policy Loss -19.3618 +trainer/Log Pis Mean 24.3367 +trainer/Log Pis Std 13.4722 +trainer/Log Pis Max 69.9919 +trainer/Log Pis Min -7.07749 +trainer/policy/mean Mean -0.0414444 +trainer/policy/mean Std 0.902109 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.83454 +trainer/policy/normal/std Std 0.675534 +trainer/policy/normal/std Max 6.01509 +trainer/policy/normal/std Min 0.290149 +trainer/policy/normal/log_std Mean 1.00067 +trainer/policy/normal/log_std Std 0.32471 +trainer/policy/normal/log_std Max 1.79427 +trainer/policy/normal/log_std Min -1.23736 +eval/num steps total 681320 +eval/num paths total 685 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.13628 +eval/Actions Std 0.941256 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59951 +time/logging (s) 0.00367334 +time/sampling batch (s) 0.534277 +time/saving (s) 0.00343735 +time/training (s) 7.56847 +time/epoch (s) 10.7094 +time/total (s) 7009.2 +Epoch -316 +---------------------------------- --------------- +2022-05-10 15:07:44.885243 PDT | [0] Epoch -315 finished +---------------------------------- --------------- +epoch -315 +replay_buffer/size 999033 +trainer/num train calls 686000 +trainer/Policy Loss -20.4228 +trainer/Log Pis Mean 24.1629 +trainer/Log Pis Std 13.5429 +trainer/Log Pis Max 70.486 +trainer/Log Pis Min -5.77573 +trainer/policy/mean Mean -0.031577 +trainer/policy/mean Std 0.908167 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81084 +trainer/policy/normal/std Std 0.687148 +trainer/policy/normal/std Max 5.74191 +trainer/policy/normal/std Min 0.234497 +trainer/policy/normal/log_std Mean 0.988777 +trainer/policy/normal/log_std Std 0.340688 +trainer/policy/normal/log_std Max 1.74779 +trainer/policy/normal/log_std Min -1.45031 +eval/num steps total 682320 +eval/num paths total 686 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.390224 +eval/Actions Std 0.86613 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55915 +time/logging (s) 0.00423549 +time/sampling batch (s) 0.532965 +time/saving (s) 0.00420175 +time/training (s) 6.93797 +time/epoch (s) 10.0385 +time/total (s) 7019.24 +Epoch -315 +---------------------------------- --------------- +2022-05-10 15:07:54.553024 PDT | [0] Epoch -314 finished +---------------------------------- --------------- +epoch -314 +replay_buffer/size 999033 +trainer/num train calls 687000 +trainer/Policy Loss -19.5205 +trainer/Log Pis Mean 25.2894 +trainer/Log Pis Std 13.0891 +trainer/Log Pis Max 63.3269 +trainer/Log Pis Min -6.60609 +trainer/policy/mean Mean -0.0398759 +trainer/policy/mean Std 0.900633 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.7644 +trainer/policy/normal/std Std 0.683732 +trainer/policy/normal/std Max 5.86841 +trainer/policy/normal/std Min 0.266714 +trainer/policy/normal/log_std Mean 0.970593 +trainer/policy/normal/log_std Std 0.347348 +trainer/policy/normal/log_std Max 1.76958 +trainer/policy/normal/log_std Min -1.32158 +eval/num steps total 683320 +eval/num paths total 687 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00195563 +eval/Actions Std 0.888509 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54113 +time/logging (s) 0.00369678 +time/sampling batch (s) 0.282393 +time/saving (s) 0.00356274 +time/training (s) 6.81349 +time/epoch (s) 9.64428 +time/total (s) 7028.89 +Epoch -314 +---------------------------------- --------------- +2022-05-10 15:08:04.755705 PDT | [0] Epoch -313 finished +---------------------------------- --------------- +epoch -313 +replay_buffer/size 999033 +trainer/num train calls 688000 +trainer/Policy Loss -20.0815 +trainer/Log Pis Mean 25.1754 +trainer/Log Pis Std 13.3004 +trainer/Log Pis Max 60.2261 +trainer/Log Pis Min -7.25822 +trainer/policy/mean Mean -0.0585438 +trainer/policy/mean Std 0.906585 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78174 +trainer/policy/normal/std Std 0.621433 +trainer/policy/normal/std Max 5.63204 +trainer/policy/normal/std Min 0.30431 +trainer/policy/normal/log_std Mean 0.989398 +trainer/policy/normal/log_std Std 0.28752 +trainer/policy/normal/log_std Max 1.72847 +trainer/policy/normal/log_std Min -1.18971 +eval/num steps total 684320 +eval/num paths total 688 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0857895 +eval/Actions Std 0.915025 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55218 +time/logging (s) 0.00375934 +time/sampling batch (s) 0.279234 +time/saving (s) 0.00341296 +time/training (s) 7.34171 +time/epoch (s) 10.1803 +time/total (s) 7039.07 +Epoch -313 +---------------------------------- --------------- +2022-05-10 15:08:14.299890 PDT | [0] Epoch -312 finished +---------------------------------- --------------- +epoch -312 +replay_buffer/size 999033 +trainer/num train calls 689000 +trainer/Policy Loss -20.6077 +trainer/Log Pis Mean 24.5683 +trainer/Log Pis Std 14.1995 +trainer/Log Pis Max 76.2036 +trainer/Log Pis Min -9.21473 +trainer/policy/mean Mean -0.0372077 +trainer/policy/mean Std 0.908332 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8779 +trainer/policy/normal/std Std 0.676726 +trainer/policy/normal/std Max 6.37094 +trainer/policy/normal/std Min 0.291261 +trainer/policy/normal/log_std Mean 1.01709 +trainer/policy/normal/log_std Std 0.319788 +trainer/policy/normal/log_std Max 1.85175 +trainer/policy/normal/log_std Min -1.23353 +eval/num steps total 685320 +eval/num paths total 689 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00304584 +eval/Actions Std 0.971045 +eval/Actions Max 0.999989 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57231 +time/logging (s) 0.00367272 +time/sampling batch (s) 0.279999 +time/saving (s) 0.00347919 +time/training (s) 6.66207 +time/epoch (s) 9.52153 +time/total (s) 7048.6 +Epoch -312 +---------------------------------- --------------- +2022-05-10 15:08:23.931130 PDT | [0] Epoch -311 finished +---------------------------------- --------------- +epoch -311 +replay_buffer/size 999033 +trainer/num train calls 690000 +trainer/Policy Loss -20.1941 +trainer/Log Pis Mean 25.4378 +trainer/Log Pis Std 13.7282 +trainer/Log Pis Max 67.39 +trainer/Log Pis Min -8.39619 +trainer/policy/mean Mean -0.0637591 +trainer/policy/mean Std 0.905652 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83818 +trainer/policy/normal/std Std 0.684185 +trainer/policy/normal/std Max 5.83509 +trainer/policy/normal/std Min 0.287548 +trainer/policy/normal/log_std Mean 1.0004 +trainer/policy/normal/log_std Std 0.332024 +trainer/policy/normal/log_std Max 1.76389 +trainer/policy/normal/log_std Min -1.24637 +eval/num steps total 686320 +eval/num paths total 690 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104641 +eval/Actions Std 0.913918 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54456 +time/logging (s) 0.00373012 +time/sampling batch (s) 0.281688 +time/saving (s) 0.00345933 +time/training (s) 6.77512 +time/epoch (s) 9.60856 +time/total (s) 7058.21 +Epoch -311 +---------------------------------- --------------- +2022-05-10 15:08:33.625279 PDT | [0] Epoch -310 finished +---------------------------------- --------------- +epoch -310 +replay_buffer/size 999033 +trainer/num train calls 691000 +trainer/Policy Loss -19.8961 +trainer/Log Pis Mean 25.7586 +trainer/Log Pis Std 13.2294 +trainer/Log Pis Max 70.5793 +trainer/Log Pis Min -6.43359 +trainer/policy/mean Mean -0.0297791 +trainer/policy/mean Std 0.909891 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83656 +trainer/policy/normal/std Std 0.660645 +trainer/policy/normal/std Max 5.87022 +trainer/policy/normal/std Min 0.313862 +trainer/policy/normal/log_std Mean 1.00224 +trainer/policy/normal/log_std Std 0.32295 +trainer/policy/normal/log_std Max 1.76989 +trainer/policy/normal/log_std Min -1.1588 +eval/num steps total 687320 +eval/num paths total 691 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101024 +eval/Actions Std 0.88674 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73967 +time/logging (s) 0.00412499 +time/sampling batch (s) 0.278591 +time/saving (s) 0.00411306 +time/training (s) 6.64542 +time/epoch (s) 9.67192 +time/total (s) 7067.88 +Epoch -310 +---------------------------------- --------------- +2022-05-10 15:08:43.467042 PDT | [0] Epoch -309 finished +---------------------------------- --------------- +epoch -309 +replay_buffer/size 999033 +trainer/num train calls 692000 +trainer/Policy Loss -19.3855 +trainer/Log Pis Mean 23.3452 +trainer/Log Pis Std 13.3022 +trainer/Log Pis Max 79.6738 +trainer/Log Pis Min -5.93641 +trainer/policy/mean Mean -0.0236625 +trainer/policy/mean Std 0.90334 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76267 +trainer/policy/normal/std Std 0.67418 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.288129 +trainer/policy/normal/log_std Mean 0.972818 +trainer/policy/normal/log_std Std 0.333572 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24435 +eval/num steps total 688320 +eval/num paths total 692 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.289742 +eval/Actions Std 0.851767 +eval/Actions Max 0.999991 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63228 +time/logging (s) 0.00382127 +time/sampling batch (s) 0.279093 +time/saving (s) 0.00364505 +time/training (s) 6.89963 +time/epoch (s) 9.81847 +time/total (s) 7077.7 +Epoch -309 +---------------------------------- --------------- +2022-05-10 15:08:53.197492 PDT | [0] Epoch -308 finished +---------------------------------- --------------- +epoch -308 +replay_buffer/size 999033 +trainer/num train calls 693000 +trainer/Policy Loss -18.7141 +trainer/Log Pis Mean 23.9343 +trainer/Log Pis Std 13.5019 +trainer/Log Pis Max 79.7902 +trainer/Log Pis Min -10.3604 +trainer/policy/mean Mean -0.0278687 +trainer/policy/mean Std 0.908253 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.74217 +trainer/policy/normal/std Std 0.639348 +trainer/policy/normal/std Max 5.38619 +trainer/policy/normal/std Min 0.322035 +trainer/policy/normal/log_std Mean 0.968703 +trainer/policy/normal/log_std Std 0.321976 +trainer/policy/normal/log_std Max 1.68384 +trainer/policy/normal/log_std Min -1.13309 +eval/num steps total 689320 +eval/num paths total 693 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128179 +eval/Actions Std 0.880319 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59287 +time/logging (s) 0.0037283 +time/sampling batch (s) 0.280715 +time/saving (s) 0.00350284 +time/training (s) 6.82686 +time/epoch (s) 9.70767 +time/total (s) 7087.41 +Epoch -308 +---------------------------------- --------------- +2022-05-10 15:09:05.419971 PDT | [0] Epoch -307 finished +---------------------------------- --------------- +epoch -307 +replay_buffer/size 999033 +trainer/num train calls 694000 +trainer/Policy Loss -19.8311 +trainer/Log Pis Mean 23.8461 +trainer/Log Pis Std 13.0504 +trainer/Log Pis Max 68.5062 +trainer/Log Pis Min -10.7538 +trainer/policy/mean Mean -0.0386436 +trainer/policy/mean Std 0.90653 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.84464 +trainer/policy/normal/std Std 0.662446 +trainer/policy/normal/std Max 5.61186 +trainer/policy/normal/std Min 0.315267 +trainer/policy/normal/log_std Mean 1.00522 +trainer/policy/normal/log_std Std 0.323185 +trainer/policy/normal/log_std Max 1.72488 +trainer/policy/normal/log_std Min -1.15434 +eval/num steps total 690320 +eval/num paths total 694 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.107996 +eval/Actions Std 0.888701 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78113 +time/logging (s) 0.00366796 +time/sampling batch (s) 0.281216 +time/saving (s) 0.00338512 +time/training (s) 9.13022 +time/epoch (s) 12.1996 +time/total (s) 7099.62 +Epoch -307 +---------------------------------- --------------- +2022-05-10 15:09:15.890462 PDT | [0] Epoch -306 finished +---------------------------------- --------------- +epoch -306 +replay_buffer/size 999033 +trainer/num train calls 695000 +trainer/Policy Loss -19.5569 +trainer/Log Pis Mean 23.9844 +trainer/Log Pis Std 13.6916 +trainer/Log Pis Max 73.978 +trainer/Log Pis Min -8.82568 +trainer/policy/mean Mean -0.0534156 +trainer/policy/mean Std 0.903657 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.72916 +trainer/policy/normal/std Std 0.652673 +trainer/policy/normal/std Max 5.41452 +trainer/policy/normal/std Min 0.253323 +trainer/policy/normal/log_std Mean 0.962219 +trainer/policy/normal/log_std Std 0.327679 +trainer/policy/normal/log_std Max 1.68908 +trainer/policy/normal/log_std Min -1.37309 +eval/num steps total 691320 +eval/num paths total 695 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00397853 +eval/Actions Std 0.898519 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57092 +time/logging (s) 0.00374091 +time/sampling batch (s) 0.282224 +time/saving (s) 0.00347205 +time/training (s) 7.58736 +time/epoch (s) 10.4477 +time/total (s) 7110.07 +Epoch -306 +---------------------------------- --------------- +2022-05-10 15:09:26.451258 PDT | [0] Epoch -305 finished +---------------------------------- --------------- +epoch -305 +replay_buffer/size 999033 +trainer/num train calls 696000 +trainer/Policy Loss -20.1586 +trainer/Log Pis Mean 25.0182 +trainer/Log Pis Std 12.8748 +trainer/Log Pis Max 66.1871 +trainer/Log Pis Min -5.02128 +trainer/policy/mean Mean -0.0229613 +trainer/policy/mean Std 0.910361 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74011 +trainer/policy/normal/std Std 0.649292 +trainer/policy/normal/std Max 5.96923 +trainer/policy/normal/std Min 0.256077 +trainer/policy/normal/log_std Mean 0.967225 +trainer/policy/normal/log_std Std 0.323429 +trainer/policy/normal/log_std Max 1.78662 +trainer/policy/normal/log_std Min -1.36228 +eval/num steps total 692320 +eval/num paths total 696 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109682 +eval/Actions Std 0.909521 +eval/Actions Max 0.999991 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55481 +time/logging (s) 0.0041936 +time/sampling batch (s) 0.532926 +time/saving (s) 0.00408105 +time/training (s) 7.44227 +time/epoch (s) 10.5383 +time/total (s) 7120.61 +Epoch -305 +---------------------------------- --------------- +2022-05-10 15:09:36.945643 PDT | [0] Epoch -304 finished +---------------------------------- --------------- +epoch -304 +replay_buffer/size 999033 +trainer/num train calls 697000 +trainer/Policy Loss -19.1869 +trainer/Log Pis Mean 24.4844 +trainer/Log Pis Std 12.6289 +trainer/Log Pis Max 71.5818 +trainer/Log Pis Min -6.79167 +trainer/policy/mean Mean -0.0439535 +trainer/policy/mean Std 0.903842 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.77945 +trainer/policy/normal/std Std 0.66818 +trainer/policy/normal/std Max 5.63984 +trainer/policy/normal/std Min 0.291076 +trainer/policy/normal/log_std Mean 0.979054 +trainer/policy/normal/log_std Std 0.335728 +trainer/policy/normal/log_std Max 1.72985 +trainer/policy/normal/log_std Min -1.23417 +eval/num steps total 693320 +eval/num paths total 697 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0409991 +eval/Actions Std 0.911844 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60798 +time/logging (s) 0.00379345 +time/sampling batch (s) 0.533283 +time/saving (s) 0.00366567 +time/training (s) 7.32174 +time/epoch (s) 10.4705 +time/total (s) 7131.08 +Epoch -304 +---------------------------------- --------------- +2022-05-10 15:09:47.671845 PDT | [0] Epoch -303 finished +---------------------------------- --------------- +epoch -303 +replay_buffer/size 999033 +trainer/num train calls 698000 +trainer/Policy Loss -19.3149 +trainer/Log Pis Mean 24.6162 +trainer/Log Pis Std 13.0343 +trainer/Log Pis Max 64.6473 +trainer/Log Pis Min -4.3313 +trainer/policy/mean Mean -0.0435816 +trainer/policy/mean Std 0.904844 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.76785 +trainer/policy/normal/std Std 0.657669 +trainer/policy/normal/std Max 5.62922 +trainer/policy/normal/std Min 0.290509 +trainer/policy/normal/log_std Mean 0.976062 +trainer/policy/normal/log_std Std 0.330372 +trainer/policy/normal/log_std Max 1.72797 +trainer/policy/normal/log_std Min -1.23612 +eval/num steps total 694320 +eval/num paths total 698 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.206395 +eval/Actions Std 0.954783 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64057 +time/logging (s) 0.0043494 +time/sampling batch (s) 0.283588 +time/saving (s) 0.00343024 +time/training (s) 7.77178 +time/epoch (s) 10.7037 +time/total (s) 7141.79 +Epoch -303 +---------------------------------- --------------- +2022-05-10 15:09:58.703451 PDT | [0] Epoch -302 finished +---------------------------------- --------------- +epoch -302 +replay_buffer/size 999033 +trainer/num train calls 699000 +trainer/Policy Loss -21.5893 +trainer/Log Pis Mean 25.4342 +trainer/Log Pis Std 13.8823 +trainer/Log Pis Max 83.2871 +trainer/Log Pis Min -10.3973 +trainer/policy/mean Mean -0.055776 +trainer/policy/mean Std 0.913191 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83703 +trainer/policy/normal/std Std 0.67932 +trainer/policy/normal/std Max 6.1717 +trainer/policy/normal/std Min 0.305605 +trainer/policy/normal/log_std Mean 1.00133 +trainer/policy/normal/log_std Std 0.325963 +trainer/policy/normal/log_std Max 1.81997 +trainer/policy/normal/log_std Min -1.18546 +eval/num steps total 695320 +eval/num paths total 699 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.27218 +eval/Actions Std 0.856365 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68842 +time/logging (s) 0.00380374 +time/sampling batch (s) 0.281541 +time/saving (s) 0.00346816 +time/training (s) 8.03042 +time/epoch (s) 11.0076 +time/total (s) 7152.8 +Epoch -302 +---------------------------------- --------------- +2022-05-10 15:10:08.335368 PDT | [0] Epoch -301 finished +---------------------------------- --------------- +epoch -301 +replay_buffer/size 999033 +trainer/num train calls 700000 +trainer/Policy Loss -19.4266 +trainer/Log Pis Mean 24.2682 +trainer/Log Pis Std 13.6346 +trainer/Log Pis Max 67.8392 +trainer/Log Pis Min -4.87058 +trainer/policy/mean Mean -0.034313 +trainer/policy/mean Std 0.902019 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.85389 +trainer/policy/normal/std Std 0.676907 +trainer/policy/normal/std Max 6.53605 +trainer/policy/normal/std Min 0.228488 +trainer/policy/normal/log_std Mean 1.00694 +trainer/policy/normal/log_std Std 0.329703 +trainer/policy/normal/log_std Max 1.87733 +trainer/policy/normal/log_std Min -1.47627 +eval/num steps total 696320 +eval/num paths total 700 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00800865 +eval/Actions Std 0.951191 +eval/Actions Max 0.99999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61797 +time/logging (s) 0.00375309 +time/sampling batch (s) 0.284031 +time/saving (s) 0.00640847 +time/training (s) 6.69645 +time/epoch (s) 9.60861 +time/total (s) 7162.41 +Epoch -301 +---------------------------------- --------------- +2022-05-10 15:10:18.570650 PDT | [0] Epoch -300 finished +---------------------------------- --------------- +epoch -300 +replay_buffer/size 999033 +trainer/num train calls 701000 +trainer/Policy Loss -19.5643 +trainer/Log Pis Mean 22.8754 +trainer/Log Pis Std 12.7966 +trainer/Log Pis Max 69.5893 +trainer/Log Pis Min -8.0385 +trainer/policy/mean Mean -0.0230787 +trainer/policy/mean Std 0.902969 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.80167 +trainer/policy/normal/std Std 0.667412 +trainer/policy/normal/std Max 6.27993 +trainer/policy/normal/std Min 0.312102 +trainer/policy/normal/log_std Mean 0.989336 +trainer/policy/normal/log_std Std 0.323173 +trainer/policy/normal/log_std Max 1.83736 +trainer/policy/normal/log_std Min -1.16443 +eval/num steps total 697320 +eval/num paths total 701 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0259799 +eval/Actions Std 0.914318 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57969 +time/logging (s) 0.00399322 +time/sampling batch (s) 0.282965 +time/saving (s) 0.00339738 +time/training (s) 7.3427 +time/epoch (s) 10.2127 +time/total (s) 7172.63 +Epoch -300 +---------------------------------- --------------- +2022-05-10 15:10:29.831143 PDT | [0] Epoch -299 finished +---------------------------------- --------------- +epoch -299 +replay_buffer/size 999033 +trainer/num train calls 702000 +trainer/Policy Loss -20.3227 +trainer/Log Pis Mean 24.9225 +trainer/Log Pis Std 12.8226 +trainer/Log Pis Max 68.2473 +trainer/Log Pis Min -4.21838 +trainer/policy/mean Mean -0.0543466 +trainer/policy/mean Std 0.908725 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.82631 +trainer/policy/normal/std Std 0.663872 +trainer/policy/normal/std Max 5.55455 +trainer/policy/normal/std Min 0.284461 +trainer/policy/normal/log_std Mean 0.997911 +trainer/policy/normal/log_std Std 0.326981 +trainer/policy/normal/log_std Max 1.71462 +trainer/policy/normal/log_std Min -1.25716 +eval/num steps total 698320 +eval/num paths total 702 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110582 +eval/Actions Std 0.914187 +eval/Actions Max 0.999992 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69858 +time/logging (s) 0.00369836 +time/sampling batch (s) 0.285724 +time/saving (s) 0.00345516 +time/training (s) 8.24586 +time/epoch (s) 11.2373 +time/total (s) 7183.87 +Epoch -299 +---------------------------------- --------------- +2022-05-10 15:10:39.576687 PDT | [0] Epoch -298 finished +---------------------------------- --------------- +epoch -298 +replay_buffer/size 999033 +trainer/num train calls 703000 +trainer/Policy Loss -19.1099 +trainer/Log Pis Mean 24.5887 +trainer/Log Pis Std 13.3056 +trainer/Log Pis Max 63.9624 +trainer/Log Pis Min -11.259 +trainer/policy/mean Mean -0.0194034 +trainer/policy/mean Std 0.906339 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77102 +trainer/policy/normal/std Std 0.650077 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.248521 +trainer/policy/normal/log_std Mean 0.980154 +trainer/policy/normal/log_std Std 0.315652 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.39223 +eval/num steps total 699320 +eval/num paths total 703 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104379 +eval/Actions Std 0.915769 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67269 +time/logging (s) 0.00366525 +time/sampling batch (s) 0.282152 +time/saving (s) 0.00341267 +time/training (s) 6.76074 +time/epoch (s) 9.72265 +time/total (s) 7193.59 +Epoch -298 +---------------------------------- --------------- +2022-05-10 15:10:49.866791 PDT | [0] Epoch -297 finished +---------------------------------- --------------- +epoch -297 +replay_buffer/size 999033 +trainer/num train calls 704000 +trainer/Policy Loss -19.875 +trainer/Log Pis Mean 26.2746 +trainer/Log Pis Std 14.1042 +trainer/Log Pis Max 69.8334 +trainer/Log Pis Min -8.71794 +trainer/policy/mean Mean -0.0487961 +trainer/policy/mean Std 0.907328 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8551 +trainer/policy/normal/std Std 0.690349 +trainer/policy/normal/std Max 6.17255 +trainer/policy/normal/std Min 0.313409 +trainer/policy/normal/log_std Mean 1.00622 +trainer/policy/normal/log_std Std 0.332206 +trainer/policy/normal/log_std Max 1.82011 +trainer/policy/normal/log_std Min -1.16025 +eval/num steps total 700320 +eval/num paths total 704 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.305574 +eval/Actions Std 0.874179 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60718 +time/logging (s) 0.00407729 +time/sampling batch (s) 0.528781 +time/saving (s) 0.00400576 +time/training (s) 7.12396 +time/epoch (s) 10.268 +time/total (s) 7203.86 +Epoch -297 +---------------------------------- --------------- +2022-05-10 15:11:00.445823 PDT | [0] Epoch -296 finished +---------------------------------- --------------- +epoch -296 +replay_buffer/size 999033 +trainer/num train calls 705000 +trainer/Policy Loss -19.256 +trainer/Log Pis Mean 24.8533 +trainer/Log Pis Std 13.164 +trainer/Log Pis Max 65.9285 +trainer/Log Pis Min -7.7155 +trainer/policy/mean Mean -0.0392454 +trainer/policy/mean Std 0.906886 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78205 +trainer/policy/normal/std Std 0.672694 +trainer/policy/normal/std Max 6.05449 +trainer/policy/normal/std Min 0.241173 +trainer/policy/normal/log_std Mean 0.97909 +trainer/policy/normal/log_std Std 0.340396 +trainer/policy/normal/log_std Max 1.8008 +trainer/policy/normal/log_std Min -1.42224 +eval/num steps total 701320 +eval/num paths total 705 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.116945 +eval/Actions Std 0.903905 +eval/Actions Max 0.999991 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62349 +time/logging (s) 0.0037549 +time/sampling batch (s) 0.781223 +time/saving (s) 0.00358808 +time/training (s) 7.14347 +time/epoch (s) 10.5555 +time/total (s) 7214.42 +Epoch -296 +---------------------------------- --------------- +2022-05-10 15:11:09.715790 PDT | [0] Epoch -295 finished +---------------------------------- --------------- +epoch -295 +replay_buffer/size 999033 +trainer/num train calls 706000 +trainer/Policy Loss -20.8682 +trainer/Log Pis Mean 24.5144 +trainer/Log Pis Std 14.0447 +trainer/Log Pis Max 72.0132 +trainer/Log Pis Min -14.0883 +trainer/policy/mean Mean -0.039934 +trainer/policy/mean Std 0.914675 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78651 +trainer/policy/normal/std Std 0.634185 +trainer/policy/normal/std Max 6.07731 +trainer/policy/normal/std Min 0.260749 +trainer/policy/normal/log_std Mean 0.987622 +trainer/policy/normal/log_std Std 0.307379 +trainer/policy/normal/log_std Max 1.80456 +trainer/policy/normal/log_std Min -1.3442 +eval/num steps total 702320 +eval/num paths total 706 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.232151 +eval/Actions Std 0.952174 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59779 +time/logging (s) 0.00370358 +time/sampling batch (s) 0.283771 +time/saving (s) 0.00342461 +time/training (s) 6.35828 +time/epoch (s) 9.24697 +time/total (s) 7223.67 +Epoch -295 +---------------------------------- --------------- +2022-05-10 15:11:19.331104 PDT | [0] Epoch -294 finished +---------------------------------- ---------------- +epoch -294 +replay_buffer/size 999033 +trainer/num train calls 707000 +trainer/Policy Loss -20.7318 +trainer/Log Pis Mean 24.5726 +trainer/Log Pis Std 13.5016 +trainer/Log Pis Max 79.6004 +trainer/Log Pis Min -11.6467 +trainer/policy/mean Mean -0.050468 +trainer/policy/mean Std 0.91436 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78042 +trainer/policy/normal/std Std 0.664403 +trainer/policy/normal/std Max 5.81808 +trainer/policy/normal/std Min 0.250858 +trainer/policy/normal/log_std Mean 0.978999 +trainer/policy/normal/log_std Std 0.340147 +trainer/policy/normal/log_std Max 1.76097 +trainer/policy/normal/log_std Min -1.38287 +eval/num steps total 703320 +eval/num paths total 707 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.000217865 +eval/Actions Std 0.960076 +eval/Actions Max 0.999975 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44948 +time/logging (s) 0.00368 +time/sampling batch (s) 0.276951 +time/saving (s) 0.00350212 +time/training (s) 6.85911 +time/epoch (s) 9.59272 +time/total (s) 7233.27 +Epoch -294 +---------------------------------- ---------------- +2022-05-10 15:11:29.728555 PDT | [0] Epoch -293 finished +---------------------------------- --------------- +epoch -293 +replay_buffer/size 999033 +trainer/num train calls 708000 +trainer/Policy Loss -20.4397 +trainer/Log Pis Mean 24.036 +trainer/Log Pis Std 12.8359 +trainer/Log Pis Max 65.3102 +trainer/Log Pis Min -7.39573 +trainer/policy/mean Mean -0.0449742 +trainer/policy/mean Std 0.907341 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82912 +trainer/policy/normal/std Std 0.670544 +trainer/policy/normal/std Max 6.43757 +trainer/policy/normal/std Min 0.330566 +trainer/policy/normal/log_std Mean 1.00097 +trainer/policy/normal/log_std Std 0.31171 +trainer/policy/normal/log_std Max 1.86215 +trainer/policy/normal/log_std Min -1.10695 +eval/num steps total 704320 +eval/num paths total 708 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.271446 +eval/Actions Std 0.819569 +eval/Actions Max 0.999992 +eval/Actions Min -0.999978 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42724 +time/logging (s) 0.00377808 +time/sampling batch (s) 0.279849 +time/saving (s) 0.00345583 +time/training (s) 7.66042 +time/epoch (s) 10.3747 +time/total (s) 7243.65 +Epoch -293 +---------------------------------- --------------- +2022-05-10 15:11:39.164231 PDT | [0] Epoch -292 finished +---------------------------------- --------------- +epoch -292 +replay_buffer/size 999033 +trainer/num train calls 709000 +trainer/Policy Loss -19.4084 +trainer/Log Pis Mean 24.657 +trainer/Log Pis Std 12.9148 +trainer/Log Pis Max 70.0231 +trainer/Log Pis Min -8.53182 +trainer/policy/mean Mean -0.0519167 +trainer/policy/mean Std 0.905796 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77885 +trainer/policy/normal/std Std 0.640109 +trainer/policy/normal/std Max 7.14765 +trainer/policy/normal/std Min 0.319592 +trainer/policy/normal/log_std Mean 0.983825 +trainer/policy/normal/log_std Std 0.312732 +trainer/policy/normal/log_std Max 1.96678 +trainer/policy/normal/log_std Min -1.14071 +eval/num steps total 705320 +eval/num paths total 709 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.166672 +eval/Actions Std 0.895462 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4739 +time/logging (s) 0.00384052 +time/sampling batch (s) 0.276829 +time/saving (s) 0.003465 +time/training (s) 6.65514 +time/epoch (s) 9.41317 +time/total (s) 7253.06 +Epoch -292 +---------------------------------- --------------- +2022-05-10 15:11:48.981843 PDT | [0] Epoch -291 finished +---------------------------------- --------------- +epoch -291 +replay_buffer/size 999033 +trainer/num train calls 710000 +trainer/Policy Loss -20.0933 +trainer/Log Pis Mean 24.7245 +trainer/Log Pis Std 13.3561 +trainer/Log Pis Max 68.0579 +trainer/Log Pis Min -8.62592 +trainer/policy/mean Mean -0.0326083 +trainer/policy/mean Std 0.904051 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76835 +trainer/policy/normal/std Std 0.641947 +trainer/policy/normal/std Max 6.43233 +trainer/policy/normal/std Min 0.305188 +trainer/policy/normal/log_std Mean 0.980141 +trainer/policy/normal/log_std Std 0.309661 +trainer/policy/normal/log_std Max 1.86134 +trainer/policy/normal/log_std Min -1.18683 +eval/num steps total 706320 +eval/num paths total 710 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132016 +eval/Actions Std 0.825916 +eval/Actions Max 0.999979 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66412 +time/logging (s) 0.00374865 +time/sampling batch (s) 0.280668 +time/saving (s) 0.00345242 +time/training (s) 6.84271 +time/epoch (s) 9.7947 +time/total (s) 7262.86 +Epoch -291 +---------------------------------- --------------- +2022-05-10 15:11:59.136601 PDT | [0] Epoch -290 finished +---------------------------------- --------------- +epoch -290 +replay_buffer/size 999033 +trainer/num train calls 711000 +trainer/Policy Loss -19.6153 +trainer/Log Pis Mean 24.1343 +trainer/Log Pis Std 13.1066 +trainer/Log Pis Max 63.1737 +trainer/Log Pis Min -9.18798 +trainer/policy/mean Mean -0.0352826 +trainer/policy/mean Std 0.906969 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.81189 +trainer/policy/normal/std Std 0.669262 +trainer/policy/normal/std Max 5.63335 +trainer/policy/normal/std Min 0.258436 +trainer/policy/normal/log_std Mean 0.992341 +trainer/policy/normal/log_std Std 0.327794 +trainer/policy/normal/log_std Max 1.7287 +trainer/policy/normal/log_std Min -1.35311 +eval/num steps total 707320 +eval/num paths total 711 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0404179 +eval/Actions Std 0.899599 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77424 +time/logging (s) 0.00367812 +time/sampling batch (s) 0.279945 +time/saving (s) 0.00342142 +time/training (s) 7.07066 +time/epoch (s) 10.1319 +time/total (s) 7272.99 +Epoch -290 +---------------------------------- --------------- +2022-05-10 15:12:09.049293 PDT | [0] Epoch -289 finished +---------------------------------- --------------- +epoch -289 +replay_buffer/size 999033 +trainer/num train calls 712000 +trainer/Policy Loss -20.6914 +trainer/Log Pis Mean 25.2411 +trainer/Log Pis Std 13.9903 +trainer/Log Pis Max 71.9826 +trainer/Log Pis Min -8.198 +trainer/policy/mean Mean -0.0336891 +trainer/policy/mean Std 0.907043 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78061 +trainer/policy/normal/std Std 0.663486 +trainer/policy/normal/std Max 6.12856 +trainer/policy/normal/std Min 0.293558 +trainer/policy/normal/log_std Mean 0.980749 +trainer/policy/normal/log_std Std 0.329206 +trainer/policy/normal/log_std Max 1.81296 +trainer/policy/normal/log_std Min -1.22568 +eval/num steps total 708320 +eval/num paths total 712 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0953612 +eval/Actions Std 0.906127 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46259 +time/logging (s) 0.00370732 +time/sampling batch (s) 0.280218 +time/saving (s) 0.00346808 +time/training (s) 7.14009 +time/epoch (s) 9.89008 +time/total (s) 7282.89 +Epoch -289 +---------------------------------- --------------- +2022-05-10 15:12:18.091126 PDT | [0] Epoch -288 finished +---------------------------------- --------------- +epoch -288 +replay_buffer/size 999033 +trainer/num train calls 713000 +trainer/Policy Loss -20.3735 +trainer/Log Pis Mean 25.0743 +trainer/Log Pis Std 13.6896 +trainer/Log Pis Max 79.7585 +trainer/Log Pis Min -6.36369 +trainer/policy/mean Mean -0.0539476 +trainer/policy/mean Std 0.906793 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76136 +trainer/policy/normal/std Std 0.651276 +trainer/policy/normal/std Max 5.32189 +trainer/policy/normal/std Min 0.331332 +trainer/policy/normal/log_std Mean 0.975564 +trainer/policy/normal/log_std Std 0.32004 +trainer/policy/normal/log_std Max 1.67183 +trainer/policy/normal/log_std Min -1.10463 +eval/num steps total 709320 +eval/num paths total 713 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10733 +eval/Actions Std 0.904589 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49588 +time/logging (s) 0.00417954 +time/sampling batch (s) 0.280228 +time/saving (s) 0.00413814 +time/training (s) 6.23502 +time/epoch (s) 9.01944 +time/total (s) 7291.91 +Epoch -288 +---------------------------------- --------------- +2022-05-10 15:12:27.850481 PDT | [0] Epoch -287 finished +---------------------------------- --------------- +epoch -287 +replay_buffer/size 999033 +trainer/num train calls 714000 +trainer/Policy Loss -19.5057 +trainer/Log Pis Mean 24.9087 +trainer/Log Pis Std 13.4425 +trainer/Log Pis Max 68.1434 +trainer/Log Pis Min -4.78071 +trainer/policy/mean Mean -0.031596 +trainer/policy/mean Std 0.909789 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.84804 +trainer/policy/normal/std Std 0.677893 +trainer/policy/normal/std Max 5.56434 +trainer/policy/normal/std Min 0.330206 +trainer/policy/normal/log_std Mean 1.00482 +trainer/policy/normal/log_std Std 0.3291 +trainer/policy/normal/log_std Max 1.71638 +trainer/policy/normal/log_std Min -1.10804 +eval/num steps total 710320 +eval/num paths total 714 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0178025 +eval/Actions Std 0.970741 +eval/Actions Max 0.999994 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65113 +time/logging (s) 0.00384881 +time/sampling batch (s) 0.531162 +time/saving (s) 0.00354259 +time/training (s) 6.5458 +time/epoch (s) 9.73548 +time/total (s) 7301.65 +Epoch -287 +---------------------------------- --------------- +2022-05-10 15:12:37.571244 PDT | [0] Epoch -286 finished +---------------------------------- --------------- +epoch -286 +replay_buffer/size 999033 +trainer/num train calls 715000 +trainer/Policy Loss -19.7233 +trainer/Log Pis Mean 24.9108 +trainer/Log Pis Std 13.2476 +trainer/Log Pis Max 77.1569 +trainer/Log Pis Min -7.8866 +trainer/policy/mean Mean -0.0554112 +trainer/policy/mean Std 0.907365 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79036 +trainer/policy/normal/std Std 0.654261 +trainer/policy/normal/std Max 5.34319 +trainer/policy/normal/std Min 0.295196 +trainer/policy/normal/log_std Mean 0.985758 +trainer/policy/normal/log_std Std 0.322524 +trainer/policy/normal/log_std Max 1.67582 +trainer/policy/normal/log_std Min -1.22012 +eval/num steps total 711320 +eval/num paths total 715 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0129042 +eval/Actions Std 0.908483 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42751 +time/logging (s) 0.00376983 +time/sampling batch (s) 0.28946 +time/saving (s) 0.00390446 +time/training (s) 6.97281 +time/epoch (s) 9.69746 +time/total (s) 7311.35 +Epoch -286 +---------------------------------- --------------- +2022-05-10 15:12:47.770579 PDT | [0] Epoch -285 finished +---------------------------------- --------------- +epoch -285 +replay_buffer/size 999033 +trainer/num train calls 716000 +trainer/Policy Loss -19.9407 +trainer/Log Pis Mean 25.5856 +trainer/Log Pis Std 13.8101 +trainer/Log Pis Max 68.8522 +trainer/Log Pis Min -7.79648 +trainer/policy/mean Mean -0.0127244 +trainer/policy/mean Std 0.911004 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85977 +trainer/policy/normal/std Std 0.659366 +trainer/policy/normal/std Max 6.31973 +trainer/policy/normal/std Min 0.29674 +trainer/policy/normal/log_std Mean 1.01197 +trainer/policy/normal/log_std Std 0.31566 +trainer/policy/normal/log_std Max 1.84368 +trainer/policy/normal/log_std Min -1.2149 +eval/num steps total 712320 +eval/num paths total 716 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0866703 +eval/Actions Std 0.910424 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70775 +time/logging (s) 0.00379908 +time/sampling batch (s) 0.279089 +time/saving (s) 0.00343963 +time/training (s) 7.18267 +time/epoch (s) 10.1767 +time/total (s) 7321.53 +Epoch -285 +---------------------------------- --------------- +2022-05-10 15:12:57.379822 PDT | [0] Epoch -284 finished +---------------------------------- --------------- +epoch -284 +replay_buffer/size 999033 +trainer/num train calls 717000 +trainer/Policy Loss -19.3048 +trainer/Log Pis Mean 24.1239 +trainer/Log Pis Std 12.8904 +trainer/Log Pis Max 66.1111 +trainer/Log Pis Min -7.19546 +trainer/policy/mean Mean -0.0140323 +trainer/policy/mean Std 0.906716 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7962 +trainer/policy/normal/std Std 0.677754 +trainer/policy/normal/std Max 5.36967 +trainer/policy/normal/std Min 0.310825 +trainer/policy/normal/log_std Mean 0.985021 +trainer/policy/normal/log_std Std 0.334683 +trainer/policy/normal/log_std Max 1.68077 +trainer/policy/normal/log_std Min -1.16853 +eval/num steps total 713320 +eval/num paths total 717 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141407 +eval/Actions Std 0.895277 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48623 +time/logging (s) 0.00369583 +time/sampling batch (s) 0.278025 +time/saving (s) 0.00338729 +time/training (s) 6.81522 +time/epoch (s) 9.58655 +time/total (s) 7331.12 +Epoch -284 +---------------------------------- --------------- +2022-05-10 15:13:08.383045 PDT | [0] Epoch -283 finished +---------------------------------- --------------- +epoch -283 +replay_buffer/size 999033 +trainer/num train calls 718000 +trainer/Policy Loss -20.4687 +trainer/Log Pis Mean 24.6815 +trainer/Log Pis Std 13.1374 +trainer/Log Pis Max 68.407 +trainer/Log Pis Min -6.79878 +trainer/policy/mean Mean -0.0343678 +trainer/policy/mean Std 0.910795 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.80168 +trainer/policy/normal/std Std 0.625112 +trainer/policy/normal/std Max 4.9884 +trainer/policy/normal/std Min 0.246615 +trainer/policy/normal/log_std Mean 0.99341 +trainer/policy/normal/log_std Std 0.308011 +trainer/policy/normal/log_std Max 1.60712 +trainer/policy/normal/log_std Min -1.39993 +eval/num steps total 714320 +eval/num paths total 718 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.202358 +eval/Actions Std 0.906844 +eval/Actions Max 0.999985 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.3757 +time/logging (s) 0.00370397 +time/sampling batch (s) 0.534461 +time/saving (s) 0.0034597 +time/training (s) 8.06298 +time/epoch (s) 10.9803 +time/total (s) 7342.1 +Epoch -283 +---------------------------------- --------------- +2022-05-10 15:13:18.588943 PDT | [0] Epoch -282 finished +---------------------------------- --------------- +epoch -282 +replay_buffer/size 999033 +trainer/num train calls 719000 +trainer/Policy Loss -19.8984 +trainer/Log Pis Mean 24.5518 +trainer/Log Pis Std 12.8204 +trainer/Log Pis Max 65.7007 +trainer/Log Pis Min -5.16594 +trainer/policy/mean Mean -0.0486009 +trainer/policy/mean Std 0.908259 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80111 +trainer/policy/normal/std Std 0.681448 +trainer/policy/normal/std Max 5.72193 +trainer/policy/normal/std Min 0.259506 +trainer/policy/normal/log_std Mean 0.985941 +trainer/policy/normal/log_std Std 0.33922 +trainer/policy/normal/log_std Max 1.74431 +trainer/policy/normal/log_std Min -1.34897 +eval/num steps total 715320 +eval/num paths total 719 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.180796 +eval/Actions Std 0.944523 +eval/Actions Max 0.999987 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55477 +time/logging (s) 0.00441813 +time/sampling batch (s) 0.532145 +time/saving (s) 0.00414894 +time/training (s) 7.08819 +time/epoch (s) 10.1837 +time/total (s) 7352.29 +Epoch -282 +---------------------------------- --------------- +2022-05-10 15:13:28.078873 PDT | [0] Epoch -281 finished +---------------------------------- --------------- +epoch -281 +replay_buffer/size 999033 +trainer/num train calls 720000 +trainer/Policy Loss -19.2916 +trainer/Log Pis Mean 25.2422 +trainer/Log Pis Std 13.1823 +trainer/Log Pis Max 69.8956 +trainer/Log Pis Min -12.332 +trainer/policy/mean Mean -0.0250517 +trainer/policy/mean Std 0.909802 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81552 +trainer/policy/normal/std Std 0.658025 +trainer/policy/normal/std Max 5.87008 +trainer/policy/normal/std Min 0.326756 +trainer/policy/normal/log_std Mean 0.994861 +trainer/policy/normal/log_std Std 0.323013 +trainer/policy/normal/log_std Max 1.76987 +trainer/policy/normal/log_std Min -1.11854 +eval/num steps total 716320 +eval/num paths total 720 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0996714 +eval/Actions Std 0.902 +eval/Actions Max 0.999999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85431 +time/logging (s) 0.00381707 +time/sampling batch (s) 0.532891 +time/saving (s) 0.00345761 +time/training (s) 6.07069 +time/epoch (s) 9.46516 +time/total (s) 7361.75 +Epoch -281 +---------------------------------- --------------- +2022-05-10 15:13:37.676927 PDT | [0] Epoch -280 finished +---------------------------------- --------------- +epoch -280 +replay_buffer/size 999033 +trainer/num train calls 721000 +trainer/Policy Loss -19.0561 +trainer/Log Pis Mean 23.89 +trainer/Log Pis Std 13.1432 +trainer/Log Pis Max 67.0289 +trainer/Log Pis Min -9.54725 +trainer/policy/mean Mean -0.0489047 +trainer/policy/mean Std 0.904374 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.83125 +trainer/policy/normal/std Std 0.661371 +trainer/policy/normal/std Max 6.16626 +trainer/policy/normal/std Min 0.279135 +trainer/policy/normal/log_std Mean 1.00271 +trainer/policy/normal/log_std Std 0.309417 +trainer/policy/normal/log_std Max 1.81909 +trainer/policy/normal/log_std Min -1.27606 +eval/num steps total 717320 +eval/num paths total 721 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.210114 +eval/Actions Std 0.961434 +eval/Actions Max 0.999993 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.526 +time/logging (s) 0.00411592 +time/sampling batch (s) 0.282657 +time/saving (s) 0.00411648 +time/training (s) 6.75831 +time/epoch (s) 9.5752 +time/total (s) 7371.33 +Epoch -280 +---------------------------------- --------------- +2022-05-10 15:13:47.389617 PDT | [0] Epoch -279 finished +---------------------------------- --------------- +epoch -279 +replay_buffer/size 999033 +trainer/num train calls 722000 +trainer/Policy Loss -18.7857 +trainer/Log Pis Mean 23.8255 +trainer/Log Pis Std 13.2497 +trainer/Log Pis Max 71.798 +trainer/Log Pis Min -11.0548 +trainer/policy/mean Mean -0.0124178 +trainer/policy/mean Std 0.905809 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.80104 +trainer/policy/normal/std Std 0.675857 +trainer/policy/normal/std Max 6.22836 +trainer/policy/normal/std Min 0.282366 +trainer/policy/normal/log_std Mean 0.98783 +trainer/policy/normal/log_std Std 0.329157 +trainer/policy/normal/log_std Max 1.82911 +trainer/policy/normal/log_std Min -1.26455 +eval/num steps total 718320 +eval/num paths total 722 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0288046 +eval/Actions Std 0.907994 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37261 +time/logging (s) 0.00386076 +time/sampling batch (s) 0.283007 +time/saving (s) 0.00373255 +time/training (s) 7.02504 +time/epoch (s) 9.68825 +time/total (s) 7381.02 +Epoch -279 +---------------------------------- --------------- +2022-05-10 15:13:56.844891 PDT | [0] Epoch -278 finished +---------------------------------- --------------- +epoch -278 +replay_buffer/size 999033 +trainer/num train calls 723000 +trainer/Policy Loss -20.3294 +trainer/Log Pis Mean 24.9982 +trainer/Log Pis Std 13.3634 +trainer/Log Pis Max 74.1287 +trainer/Log Pis Min -6.52593 +trainer/policy/mean Mean -0.0347083 +trainer/policy/mean Std 0.910362 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.77994 +trainer/policy/normal/std Std 0.668442 +trainer/policy/normal/std Max 5.88215 +trainer/policy/normal/std Min 0.303171 +trainer/policy/normal/log_std Mean 0.979911 +trainer/policy/normal/log_std Std 0.330808 +trainer/policy/normal/log_std Max 1.77192 +trainer/policy/normal/log_std Min -1.19346 +eval/num steps total 719320 +eval/num paths total 723 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.123602 +eval/Actions Std 0.832895 +eval/Actions Max 0.999998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51934 +time/logging (s) 0.00377416 +time/sampling batch (s) 0.537008 +time/saving (s) 0.00344493 +time/training (s) 6.36815 +time/epoch (s) 9.43171 +time/total (s) 7390.46 +Epoch -278 +---------------------------------- --------------- +2022-05-10 15:14:06.490459 PDT | [0] Epoch -277 finished +---------------------------------- --------------- +epoch -277 +replay_buffer/size 999033 +trainer/num train calls 724000 +trainer/Policy Loss -20.3732 +trainer/Log Pis Mean 25.6123 +trainer/Log Pis Std 13.7341 +trainer/Log Pis Max 70.4442 +trainer/Log Pis Min -4.13298 +trainer/policy/mean Mean -0.0210354 +trainer/policy/mean Std 0.910233 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.79893 +trainer/policy/normal/std Std 0.629338 +trainer/policy/normal/std Max 5.25896 +trainer/policy/normal/std Min 0.317047 +trainer/policy/normal/log_std Mean 0.992115 +trainer/policy/normal/log_std Std 0.308725 +trainer/policy/normal/log_std Max 1.65993 +trainer/policy/normal/log_std Min -1.1487 +eval/num steps total 720320 +eval/num paths total 724 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00207548 +eval/Actions Std 0.898401 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67269 +time/logging (s) 0.00366882 +time/sampling batch (s) 0.280962 +time/saving (s) 0.00345857 +time/training (s) 6.6614 +time/epoch (s) 9.62218 +time/total (s) 7400.08 +Epoch -277 +---------------------------------- --------------- +2022-05-10 15:14:17.182837 PDT | [0] Epoch -276 finished +---------------------------------- --------------- +epoch -276 +replay_buffer/size 999033 +trainer/num train calls 725000 +trainer/Policy Loss -20.022 +trainer/Log Pis Mean 25.6634 +trainer/Log Pis Std 12.9427 +trainer/Log Pis Max 66.6607 +trainer/Log Pis Min -7.48512 +trainer/policy/mean Mean -0.0389712 +trainer/policy/mean Std 0.904942 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81187 +trainer/policy/normal/std Std 0.680104 +trainer/policy/normal/std Max 5.28865 +trainer/policy/normal/std Min 0.263505 +trainer/policy/normal/log_std Mean 0.990183 +trainer/policy/normal/log_std Std 0.337204 +trainer/policy/normal/log_std Max 1.66556 +trainer/policy/normal/log_std Min -1.33368 +eval/num steps total 721320 +eval/num paths total 725 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.362876 +eval/Actions Std 0.88017 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53238 +time/logging (s) 0.00381098 +time/sampling batch (s) 0.28151 +time/saving (s) 0.00343626 +time/training (s) 7.84837 +time/epoch (s) 10.6695 +time/total (s) 7410.76 +Epoch -276 +---------------------------------- --------------- +2022-05-10 15:14:26.349275 PDT | [0] Epoch -275 finished +---------------------------------- --------------- +epoch -275 +replay_buffer/size 999033 +trainer/num train calls 726000 +trainer/Policy Loss -19.7277 +trainer/Log Pis Mean 24.2805 +trainer/Log Pis Std 13.3967 +trainer/Log Pis Max 75.8508 +trainer/Log Pis Min -10.6806 +trainer/policy/mean Mean -0.0172246 +trainer/policy/mean Std 0.906232 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.79953 +trainer/policy/normal/std Std 0.643725 +trainer/policy/normal/std Max 6.04246 +trainer/policy/normal/std Min 0.318985 +trainer/policy/normal/log_std Mean 0.992374 +trainer/policy/normal/log_std Std 0.305254 +trainer/policy/normal/log_std Max 1.79881 +trainer/policy/normal/log_std Min -1.14261 +eval/num steps total 722320 +eval/num paths total 726 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0919962 +eval/Actions Std 0.90881 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76124 +time/logging (s) 0.00401147 +time/sampling batch (s) 0.280184 +time/saving (s) 0.00356799 +time/training (s) 6.09477 +time/epoch (s) 9.14377 +time/total (s) 7419.9 +Epoch -275 +---------------------------------- --------------- +2022-05-10 15:14:35.707596 PDT | [0] Epoch -274 finished +---------------------------------- --------------- +epoch -274 +replay_buffer/size 999033 +trainer/num train calls 727000 +trainer/Policy Loss -19.2712 +trainer/Log Pis Mean 24.459 +trainer/Log Pis Std 13.5002 +trainer/Log Pis Max 71.7768 +trainer/Log Pis Min -11.9922 +trainer/policy/mean Mean -0.0437249 +trainer/policy/mean Std 0.90963 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79694 +trainer/policy/normal/std Std 0.656698 +trainer/policy/normal/std Max 5.36589 +trainer/policy/normal/std Min 0.293016 +trainer/policy/normal/log_std Mean 0.987543 +trainer/policy/normal/log_std Std 0.326302 +trainer/policy/normal/log_std Max 1.68006 +trainer/policy/normal/log_std Min -1.22753 +eval/num steps total 723320 +eval/num paths total 727 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.291769 +eval/Actions Std 0.731176 +eval/Actions Max 0.999997 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46189 +time/logging (s) 0.00425643 +time/sampling batch (s) 0.279838 +time/saving (s) 0.00422024 +time/training (s) 6.58502 +time/epoch (s) 9.33523 +time/total (s) 7429.24 +Epoch -274 +---------------------------------- --------------- +2022-05-10 15:14:45.194823 PDT | [0] Epoch -273 finished +---------------------------------- --------------- +epoch -273 +replay_buffer/size 999033 +trainer/num train calls 728000 +trainer/Policy Loss -19.6467 +trainer/Log Pis Mean 24.211 +trainer/Log Pis Std 13.5204 +trainer/Log Pis Max 70.6453 +trainer/Log Pis Min -8.80371 +trainer/policy/mean Mean -0.0201696 +trainer/policy/mean Std 0.90743 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79329 +trainer/policy/normal/std Std 0.652282 +trainer/policy/normal/std Max 6.1329 +trainer/policy/normal/std Min 0.251449 +trainer/policy/normal/log_std Mean 0.987692 +trainer/policy/normal/log_std Std 0.319195 +trainer/policy/normal/log_std Max 1.81367 +trainer/policy/normal/log_std Min -1.38052 +eval/num steps total 724320 +eval/num paths total 728 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.192015 +eval/Actions Std 0.899449 +eval/Actions Max 0.999978 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70879 +time/logging (s) 0.00379663 +time/sampling batch (s) 0.27977 +time/saving (s) 0.00361784 +time/training (s) 6.46744 +time/epoch (s) 9.46342 +time/total (s) 7438.71 +Epoch -273 +---------------------------------- --------------- +2022-05-10 15:14:55.631589 PDT | [0] Epoch -272 finished +---------------------------------- --------------- +epoch -272 +replay_buffer/size 999033 +trainer/num train calls 729000 +trainer/Policy Loss -19.9648 +trainer/Log Pis Mean 24.246 +trainer/Log Pis Std 13.0786 +trainer/Log Pis Max 66.2254 +trainer/Log Pis Min -8.59328 +trainer/policy/mean Mean -0.0191075 +trainer/policy/mean Std 0.903966 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79582 +trainer/policy/normal/std Std 0.677783 +trainer/policy/normal/std Max 6.74353 +trainer/policy/normal/std Min 0.282399 +trainer/policy/normal/log_std Mean 0.984661 +trainer/policy/normal/log_std Std 0.336154 +trainer/policy/normal/log_std Max 1.90858 +trainer/policy/normal/log_std Min -1.26444 +eval/num steps total 725320 +eval/num paths total 729 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115372 +eval/Actions Std 0.905616 +eval/Actions Max 0.999987 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65695 +time/logging (s) 0.00372569 +time/sampling batch (s) 0.530003 +time/saving (s) 0.00340036 +time/training (s) 7.21967 +time/epoch (s) 10.4137 +time/total (s) 7449.13 +Epoch -272 +---------------------------------- --------------- +2022-05-10 15:15:05.003970 PDT | [0] Epoch -271 finished +---------------------------------- --------------- +epoch -271 +replay_buffer/size 999033 +trainer/num train calls 730000 +trainer/Policy Loss -20.0487 +trainer/Log Pis Mean 23.6148 +trainer/Log Pis Std 13.4458 +trainer/Log Pis Max 64.0417 +trainer/Log Pis Min -10.2352 +trainer/policy/mean Mean -0.0322492 +trainer/policy/mean Std 0.903409 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.75455 +trainer/policy/normal/std Std 0.668384 +trainer/policy/normal/std Max 5.36254 +trainer/policy/normal/std Min 0.263507 +trainer/policy/normal/log_std Mean 0.968792 +trainer/policy/normal/log_std Std 0.340572 +trainer/policy/normal/log_std Max 1.67944 +trainer/policy/normal/log_std Min -1.33367 +eval/num steps total 726320 +eval/num paths total 730 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.076156 +eval/Actions Std 0.8964 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39353 +time/logging (s) 0.00370857 +time/sampling batch (s) 0.280462 +time/saving (s) 0.00346717 +time/training (s) 6.66843 +time/epoch (s) 9.3496 +time/total (s) 7458.48 +Epoch -271 +---------------------------------- --------------- +2022-05-10 15:15:15.714687 PDT | [0] Epoch -270 finished +---------------------------------- --------------- +epoch -270 +replay_buffer/size 999033 +trainer/num train calls 731000 +trainer/Policy Loss -20.917 +trainer/Log Pis Mean 24.2073 +trainer/Log Pis Std 13.9407 +trainer/Log Pis Max 65.7265 +trainer/Log Pis Min -12.8879 +trainer/policy/mean Mean -0.0207653 +trainer/policy/mean Std 0.907376 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.84363 +trainer/policy/normal/std Std 0.688251 +trainer/policy/normal/std Max 5.80207 +trainer/policy/normal/std Min 0.33083 +trainer/policy/normal/log_std Mean 1.00344 +trainer/policy/normal/log_std Std 0.324509 +trainer/policy/normal/log_std Max 1.75821 +trainer/policy/normal/log_std Min -1.10615 +eval/num steps total 727320 +eval/num paths total 731 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0970287 +eval/Actions Std 0.916648 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.33265 +time/logging (s) 0.00378469 +time/sampling batch (s) 0.284236 +time/saving (s) 0.00348817 +time/training (s) 8.06353 +time/epoch (s) 10.6877 +time/total (s) 7469.17 +Epoch -270 +---------------------------------- --------------- +2022-05-10 15:15:26.577316 PDT | [0] Epoch -269 finished +---------------------------------- --------------- +epoch -269 +replay_buffer/size 999033 +trainer/num train calls 732000 +trainer/Policy Loss -19.6192 +trainer/Log Pis Mean 24.946 +trainer/Log Pis Std 13.3671 +trainer/Log Pis Max 70.571 +trainer/Log Pis Min -9.45237 +trainer/policy/mean Mean -0.0150098 +trainer/policy/mean Std 0.9083 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77118 +trainer/policy/normal/std Std 0.671627 +trainer/policy/normal/std Max 5.28609 +trainer/policy/normal/std Min 0.269603 +trainer/policy/normal/log_std Mean 0.975247 +trainer/policy/normal/log_std Std 0.33913 +trainer/policy/normal/log_std Max 1.66508 +trainer/policy/normal/log_std Min -1.3108 +eval/num steps total 728320 +eval/num paths total 732 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0811906 +eval/Actions Std 0.845935 +eval/Actions Max 0.999996 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.88713 +time/logging (s) 0.00421374 +time/sampling batch (s) 0.537495 +time/saving (s) 0.00411609 +time/training (s) 7.40675 +time/epoch (s) 10.8397 +time/total (s) 7480.01 +Epoch -269 +---------------------------------- --------------- +2022-05-10 15:15:36.132024 PDT | [0] Epoch -268 finished +---------------------------------- --------------- +epoch -268 +replay_buffer/size 999033 +trainer/num train calls 733000 +trainer/Policy Loss -18.8347 +trainer/Log Pis Mean 23.9162 +trainer/Log Pis Std 14.2945 +trainer/Log Pis Max 67.8935 +trainer/Log Pis Min -7.13314 +trainer/policy/mean Mean -0.0205647 +trainer/policy/mean Std 0.903909 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.83309 +trainer/policy/normal/std Std 0.678389 +trainer/policy/normal/std Max 6.8509 +trainer/policy/normal/std Min 0.344691 +trainer/policy/normal/log_std Mean 1.00036 +trainer/policy/normal/log_std Std 0.322485 +trainer/policy/normal/log_std Max 1.92438 +trainer/policy/normal/log_std Min -1.06511 +eval/num steps total 729320 +eval/num paths total 733 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121827 +eval/Actions Std 0.896035 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78581 +time/logging (s) 0.00384575 +time/sampling batch (s) 0.282794 +time/saving (s) 0.00370523 +time/training (s) 6.45432 +time/epoch (s) 9.53047 +time/total (s) 7489.54 +Epoch -268 +---------------------------------- --------------- +2022-05-10 15:15:47.378396 PDT | [0] Epoch -267 finished +---------------------------------- --------------- +epoch -267 +replay_buffer/size 999033 +trainer/num train calls 734000 +trainer/Policy Loss -19.3375 +trainer/Log Pis Mean 24.3319 +trainer/Log Pis Std 13.4168 +trainer/Log Pis Max 76.2476 +trainer/Log Pis Min -7.22565 +trainer/policy/mean Mean -0.0263859 +trainer/policy/mean Std 0.906344 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80118 +trainer/policy/normal/std Std 0.673687 +trainer/policy/normal/std Max 5.83047 +trainer/policy/normal/std Min 0.298077 +trainer/policy/normal/log_std Mean 0.98773 +trainer/policy/normal/log_std Std 0.330216 +trainer/policy/normal/log_std Max 1.7631 +trainer/policy/normal/log_std Min -1.2104 +eval/num steps total 730320 +eval/num paths total 734 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.34313 +eval/Actions Std 0.827928 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39439 +time/logging (s) 0.00370313 +time/sampling batch (s) 0.534576 +time/saving (s) 0.00348524 +time/training (s) 8.28662 +time/epoch (s) 11.2228 +time/total (s) 7500.77 +Epoch -267 +---------------------------------- --------------- +2022-05-10 15:15:57.899636 PDT | [0] Epoch -266 finished +---------------------------------- --------------- +epoch -266 +replay_buffer/size 999033 +trainer/num train calls 735000 +trainer/Policy Loss -19.4007 +trainer/Log Pis Mean 24.5983 +trainer/Log Pis Std 13.0455 +trainer/Log Pis Max 70.9006 +trainer/Log Pis Min -4.70652 +trainer/policy/mean Mean -0.050396 +trainer/policy/mean Std 0.906085 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80241 +trainer/policy/normal/std Std 0.64767 +trainer/policy/normal/std Max 6.91492 +trainer/policy/normal/std Min 0.300799 +trainer/policy/normal/log_std Mean 0.992861 +trainer/policy/normal/log_std Std 0.308778 +trainer/policy/normal/log_std Max 1.93368 +trainer/policy/normal/log_std Min -1.20131 +eval/num steps total 731320 +eval/num paths total 735 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0964981 +eval/Actions Std 0.917099 +eval/Actions Max 0.999994 +eval/Actions Min -0.99998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73502 +time/logging (s) 0.00370144 +time/sampling batch (s) 0.284419 +time/saving (s) 0.00342536 +time/training (s) 7.47148 +time/epoch (s) 10.498 +time/total (s) 7511.27 +Epoch -266 +---------------------------------- --------------- +2022-05-10 15:16:07.936773 PDT | [0] Epoch -265 finished +---------------------------------- --------------- +epoch -265 +replay_buffer/size 999033 +trainer/num train calls 736000 +trainer/Policy Loss -19.6396 +trainer/Log Pis Mean 24.0191 +trainer/Log Pis Std 13.0335 +trainer/Log Pis Max 65.997 +trainer/Log Pis Min -6.74462 +trainer/policy/mean Mean -0.0519427 +trainer/policy/mean Std 0.901998 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76975 +trainer/policy/normal/std Std 0.680081 +trainer/policy/normal/std Max 6.13375 +trainer/policy/normal/std Min 0.280284 +trainer/policy/normal/log_std Mean 0.974017 +trainer/policy/normal/log_std Std 0.340464 +trainer/policy/normal/log_std Max 1.81381 +trainer/policy/normal/log_std Min -1.27195 +eval/num steps total 732320 +eval/num paths total 736 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0971122 +eval/Actions Std 0.887832 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69911 +time/logging (s) 0.00370485 +time/sampling batch (s) 0.281281 +time/saving (s) 0.00343453 +time/training (s) 7.02671 +time/epoch (s) 10.0142 +time/total (s) 7521.29 +Epoch -265 +---------------------------------- --------------- +2022-05-10 15:16:18.601009 PDT | [0] Epoch -264 finished +---------------------------------- --------------- +epoch -264 +replay_buffer/size 999033 +trainer/num train calls 737000 +trainer/Policy Loss -19.8024 +trainer/Log Pis Mean 24.493 +trainer/Log Pis Std 13.0263 +trainer/Log Pis Max 59.5948 +trainer/Log Pis Min -7.08359 +trainer/policy/mean Mean -0.0207266 +trainer/policy/mean Std 0.905746 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.84658 +trainer/policy/normal/std Std 0.679607 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.307157 +trainer/policy/normal/log_std Mean 1.00523 +trainer/policy/normal/log_std Std 0.323179 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.1804 +eval/num steps total 733320 +eval/num paths total 737 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00692992 +eval/Actions Std 0.917535 +eval/Actions Max 0.999997 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76665 +time/logging (s) 0.00375036 +time/sampling batch (s) 0.535812 +time/saving (s) 0.00348927 +time/training (s) 7.33136 +time/epoch (s) 10.6411 +time/total (s) 7531.93 +Epoch -264 +---------------------------------- --------------- +2022-05-10 15:16:27.778908 PDT | [0] Epoch -263 finished +---------------------------------- --------------- +epoch -263 +replay_buffer/size 999033 +trainer/num train calls 738000 +trainer/Policy Loss -19.1973 +trainer/Log Pis Mean 23.2923 +trainer/Log Pis Std 13.2591 +trainer/Log Pis Max 64.8836 +trainer/Log Pis Min -6.21645 +trainer/policy/mean Mean -0.0248669 +trainer/policy/mean Std 0.904417 +trainer/policy/mean Max 0.999972 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77616 +trainer/policy/normal/std Std 0.68246 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.294262 +trainer/policy/normal/log_std Mean 0.975599 +trainer/policy/normal/log_std Std 0.344618 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22329 +eval/num steps total 734320 +eval/num paths total 738 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.172795 +eval/Actions Std 0.803061 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63244 +time/logging (s) 0.00394532 +time/sampling batch (s) 0.287468 +time/saving (s) 0.00404176 +time/training (s) 6.22674 +time/epoch (s) 9.15463 +time/total (s) 7541.09 +Epoch -263 +---------------------------------- --------------- +2022-05-10 15:16:37.672439 PDT | [0] Epoch -262 finished +---------------------------------- --------------- +epoch -262 +replay_buffer/size 999033 +trainer/num train calls 739000 +trainer/Policy Loss -20.0865 +trainer/Log Pis Mean 24.5509 +trainer/Log Pis Std 13.8021 +trainer/Log Pis Max 73.8724 +trainer/Log Pis Min -12.2164 +trainer/policy/mean Mean -0.0320593 +trainer/policy/mean Std 0.905395 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79657 +trainer/policy/normal/std Std 0.675178 +trainer/policy/normal/std Max 6.97958 +trainer/policy/normal/std Min 0.2345 +trainer/policy/normal/log_std Mean 0.985346 +trainer/policy/normal/log_std Std 0.334656 +trainer/policy/normal/log_std Max 1.94299 +trainer/policy/normal/log_std Min -1.4503 +eval/num steps total 735320 +eval/num paths total 739 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.079293 +eval/Actions Std 0.903485 +eval/Actions Max 0.999993 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53697 +time/logging (s) 0.00386721 +time/sampling batch (s) 0.285509 +time/saving (s) 0.00421827 +time/training (s) 7.03931 +time/epoch (s) 9.86987 +time/total (s) 7550.96 +Epoch -262 +---------------------------------- --------------- +2022-05-10 15:16:47.745100 PDT | [0] Epoch -261 finished +---------------------------------- --------------- +epoch -261 +replay_buffer/size 999033 +trainer/num train calls 740000 +trainer/Policy Loss -17.5582 +trainer/Log Pis Mean 23.7413 +trainer/Log Pis Std 13.3071 +trainer/Log Pis Max 64.0669 +trainer/Log Pis Min -6.00219 +trainer/policy/mean Mean -0.015184 +trainer/policy/mean Std 0.904401 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.73732 +trainer/policy/normal/std Std 0.64823 +trainer/policy/normal/std Max 5.20874 +trainer/policy/normal/std Min 0.408616 +trainer/policy/normal/log_std Mean 0.96811 +trainer/policy/normal/log_std Std 0.309732 +trainer/policy/normal/log_std Max 1.65034 +trainer/policy/normal/log_std Min -0.89498 +eval/num steps total 736320 +eval/num paths total 740 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.114258 +eval/Actions Std 0.92308 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5614 +time/logging (s) 0.0036994 +time/sampling batch (s) 0.281752 +time/saving (s) 0.00345159 +time/training (s) 7.19819 +time/epoch (s) 10.0485 +time/total (s) 7561.02 +Epoch -261 +---------------------------------- --------------- +2022-05-10 15:16:57.775477 PDT | [0] Epoch -260 finished +---------------------------------- --------------- +epoch -260 +replay_buffer/size 999033 +trainer/num train calls 741000 +trainer/Policy Loss -20.4113 +trainer/Log Pis Mean 24.3122 +trainer/Log Pis Std 13.8677 +trainer/Log Pis Max 78.1222 +trainer/Log Pis Min -13.2845 +trainer/policy/mean Mean -0.0394026 +trainer/policy/mean Std 0.904296 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.82623 +trainer/policy/normal/std Std 0.657856 +trainer/policy/normal/std Max 5.95541 +trainer/policy/normal/std Min 0.309445 +trainer/policy/normal/log_std Mean 1.00035 +trainer/policy/normal/log_std Std 0.31351 +trainer/policy/normal/log_std Max 1.7843 +trainer/policy/normal/log_std Min -1.17297 +eval/num steps total 737320 +eval/num paths total 741 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0868753 +eval/Actions Std 0.887857 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4149 +time/logging (s) 0.00372909 +time/sampling batch (s) 0.530185 +time/saving (s) 0.00343582 +time/training (s) 7.05532 +time/epoch (s) 10.0076 +time/total (s) 7571.03 +Epoch -260 +---------------------------------- --------------- +2022-05-10 15:17:07.883973 PDT | [0] Epoch -259 finished +---------------------------------- --------------- +epoch -259 +replay_buffer/size 999033 +trainer/num train calls 742000 +trainer/Policy Loss -20.6964 +trainer/Log Pis Mean 23.6839 +trainer/Log Pis Std 12.6464 +trainer/Log Pis Max 64.4062 +trainer/Log Pis Min -5.82553 +trainer/policy/mean Mean -0.0489515 +trainer/policy/mean Std 0.911564 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74983 +trainer/policy/normal/std Std 0.633172 +trainer/policy/normal/std Max 5.59059 +trainer/policy/normal/std Min 0.348302 +trainer/policy/normal/log_std Mean 0.974807 +trainer/policy/normal/log_std Std 0.302635 +trainer/policy/normal/log_std Max 1.72108 +trainer/policy/normal/log_std Min -1.05469 +eval/num steps total 738320 +eval/num paths total 742 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.393496 +eval/Actions Std 0.849565 +eval/Actions Max 0.999991 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71769 +time/logging (s) 0.00370973 +time/sampling batch (s) 0.2803 +time/saving (s) 0.00345247 +time/training (s) 7.08042 +time/epoch (s) 10.0856 +time/total (s) 7581.11 +Epoch -259 +---------------------------------- --------------- +2022-05-10 15:17:17.414812 PDT | [0] Epoch -258 finished +---------------------------------- --------------- +epoch -258 +replay_buffer/size 999033 +trainer/num train calls 743000 +trainer/Policy Loss -20.657 +trainer/Log Pis Mean 24.093 +trainer/Log Pis Std 12.7947 +trainer/Log Pis Max 72.589 +trainer/Log Pis Min -7.18321 +trainer/policy/mean Mean -0.0190918 +trainer/policy/mean Std 0.905775 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.792 +trainer/policy/normal/std Std 0.655721 +trainer/policy/normal/std Max 5.71196 +trainer/policy/normal/std Min 0.233456 +trainer/policy/normal/log_std Mean 0.987363 +trainer/policy/normal/log_std Std 0.316854 +trainer/policy/normal/log_std Max 1.74256 +trainer/policy/normal/log_std Min -1.45476 +eval/num steps total 739320 +eval/num paths total 743 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0300473 +eval/Actions Std 0.908241 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7243 +time/logging (s) 0.00376342 +time/sampling batch (s) 0.279818 +time/saving (s) 0.00343878 +time/training (s) 6.49635 +time/epoch (s) 9.50767 +time/total (s) 7590.62 +Epoch -258 +---------------------------------- --------------- +2022-05-10 15:17:27.776469 PDT | [0] Epoch -257 finished +---------------------------------- --------------- +epoch -257 +replay_buffer/size 999033 +trainer/num train calls 744000 +trainer/Policy Loss -18.6931 +trainer/Log Pis Mean 24.4321 +trainer/Log Pis Std 13.657 +trainer/Log Pis Max 70.1051 +trainer/Log Pis Min -5.75127 +trainer/policy/mean Mean -0.036676 +trainer/policy/mean Std 0.908544 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7773 +trainer/policy/normal/std Std 0.698246 +trainer/policy/normal/std Max 5.98527 +trainer/policy/normal/std Min 0.255506 +trainer/policy/normal/log_std Mean 0.974518 +trainer/policy/normal/log_std Std 0.349583 +trainer/policy/normal/log_std Max 1.7893 +trainer/policy/normal/log_std Min -1.36451 +eval/num steps total 740320 +eval/num paths total 744 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.067506 +eval/Actions Std 0.852956 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64679 +time/logging (s) 0.00425336 +time/sampling batch (s) 0.284523 +time/saving (s) 0.00414185 +time/training (s) 7.39921 +time/epoch (s) 10.3389 +time/total (s) 7600.97 +Epoch -257 +---------------------------------- --------------- +2022-05-10 15:17:37.397182 PDT | [0] Epoch -256 finished +---------------------------------- --------------- +epoch -256 +replay_buffer/size 999033 +trainer/num train calls 745000 +trainer/Policy Loss -20.1054 +trainer/Log Pis Mean 25.3083 +trainer/Log Pis Std 13.8354 +trainer/Log Pis Max 65.6934 +trainer/Log Pis Min -10.0336 +trainer/policy/mean Mean -0.0316182 +trainer/policy/mean Std 0.910677 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80321 +trainer/policy/normal/std Std 0.647795 +trainer/policy/normal/std Max 5.40674 +trainer/policy/normal/std Min 0.382638 +trainer/policy/normal/log_std Mean 0.993245 +trainer/policy/normal/log_std Std 0.306619 +trainer/policy/normal/log_std Max 1.68765 +trainer/policy/normal/log_std Min -0.960665 +eval/num steps total 741320 +eval/num paths total 745 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.014832 +eval/Actions Std 0.904909 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50123 +time/logging (s) 0.00390152 +time/sampling batch (s) 0.283636 +time/saving (s) 0.00365888 +time/training (s) 6.80432 +time/epoch (s) 9.59675 +time/total (s) 7610.57 +Epoch -256 +---------------------------------- --------------- +2022-05-10 15:17:48.565106 PDT | [0] Epoch -255 finished +---------------------------------- --------------- +epoch -255 +replay_buffer/size 999033 +trainer/num train calls 746000 +trainer/Policy Loss -19.9722 +trainer/Log Pis Mean 24.252 +trainer/Log Pis Std 12.8895 +trainer/Log Pis Max 65.5867 +trainer/Log Pis Min -7.31737 +trainer/policy/mean Mean -0.0414822 +trainer/policy/mean Std 0.910421 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77829 +trainer/policy/normal/std Std 0.669134 +trainer/policy/normal/std Max 5.49601 +trainer/policy/normal/std Min 0.285574 +trainer/policy/normal/log_std Mean 0.978982 +trainer/policy/normal/log_std Std 0.333136 +trainer/policy/normal/log_std Max 1.70402 +trainer/policy/normal/log_std Min -1.25325 +eval/num steps total 742320 +eval/num paths total 746 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.075994 +eval/Actions Std 0.894336 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75633 +time/logging (s) 0.00373449 +time/sampling batch (s) 0.284708 +time/saving (s) 0.00344265 +time/training (s) 8.09639 +time/epoch (s) 11.1446 +time/total (s) 7621.71 +Epoch -255 +---------------------------------- --------------- +2022-05-10 15:17:57.777258 PDT | [0] Epoch -254 finished +---------------------------------- --------------- +epoch -254 +replay_buffer/size 999033 +trainer/num train calls 747000 +trainer/Policy Loss -20.4708 +trainer/Log Pis Mean 24.3668 +trainer/Log Pis Std 13.7372 +trainer/Log Pis Max 78.8229 +trainer/Log Pis Min -6.176 +trainer/policy/mean Mean -0.0331745 +trainer/policy/mean Std 0.908918 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999979 +trainer/policy/normal/std Mean 2.81814 +trainer/policy/normal/std Std 0.648392 +trainer/policy/normal/std Max 6.28227 +trainer/policy/normal/std Min 0.279414 +trainer/policy/normal/log_std Mean 0.998138 +trainer/policy/normal/log_std Std 0.31078 +trainer/policy/normal/log_std Max 1.83773 +trainer/policy/normal/log_std Min -1.27506 +eval/num steps total 743320 +eval/num paths total 747 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.104474 +eval/Actions Std 0.92128 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40352 +time/logging (s) 0.00379347 +time/sampling batch (s) 0.279639 +time/saving (s) 0.00346525 +time/training (s) 6.499 +time/epoch (s) 9.18941 +time/total (s) 7630.91 +Epoch -254 +---------------------------------- --------------- +2022-05-10 15:18:08.070261 PDT | [0] Epoch -253 finished +---------------------------------- --------------- +epoch -253 +replay_buffer/size 999033 +trainer/num train calls 748000 +trainer/Policy Loss -20.6585 +trainer/Log Pis Mean 23.5324 +trainer/Log Pis Std 13.1392 +trainer/Log Pis Max 71.0891 +trainer/Log Pis Min -5.17766 +trainer/policy/mean Mean -0.0339391 +trainer/policy/mean Std 0.899974 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.75948 +trainer/policy/normal/std Std 0.670839 +trainer/policy/normal/std Max 5.66158 +trainer/policy/normal/std Min 0.241649 +trainer/policy/normal/log_std Mean 0.970983 +trainer/policy/normal/log_std Std 0.338286 +trainer/policy/normal/log_std Max 1.7337 +trainer/policy/normal/log_std Min -1.42027 +eval/num steps total 744320 +eval/num paths total 748 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00431586 +eval/Actions Std 0.905884 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5611 +time/logging (s) 0.00376933 +time/sampling batch (s) 0.282567 +time/saving (s) 0.00340648 +time/training (s) 7.41915 +time/epoch (s) 10.27 +time/total (s) 7641.18 +Epoch -253 +---------------------------------- --------------- +2022-05-10 15:18:17.812167 PDT | [0] Epoch -252 finished +---------------------------------- --------------- +epoch -252 +replay_buffer/size 999033 +trainer/num train calls 749000 +trainer/Policy Loss -20.1904 +trainer/Log Pis Mean 24.1757 +trainer/Log Pis Std 13.3304 +trainer/Log Pis Max 70.932 +trainer/Log Pis Min -10.0357 +trainer/policy/mean Mean -0.0467856 +trainer/policy/mean Std 0.909608 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81077 +trainer/policy/normal/std Std 0.652941 +trainer/policy/normal/std Max 5.64319 +trainer/policy/normal/std Min 0.324145 +trainer/policy/normal/log_std Mean 0.994266 +trainer/policy/normal/log_std Std 0.317653 +trainer/policy/normal/log_std Max 1.73045 +trainer/policy/normal/log_std Min -1.12656 +eval/num steps total 745320 +eval/num paths total 749 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0160373 +eval/Actions Std 0.915246 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58872 +time/logging (s) 0.00367392 +time/sampling batch (s) 0.285454 +time/saving (s) 0.00344805 +time/training (s) 6.83722 +time/epoch (s) 9.71852 +time/total (s) 7650.9 +Epoch -252 +---------------------------------- --------------- +2022-05-10 15:18:28.090801 PDT | [0] Epoch -251 finished +---------------------------------- --------------- +epoch -251 +replay_buffer/size 999033 +trainer/num train calls 750000 +trainer/Policy Loss -20.3867 +trainer/Log Pis Mean 23.7758 +trainer/Log Pis Std 13.087 +trainer/Log Pis Max 59.6896 +trainer/Log Pis Min -6.01633 +trainer/policy/mean Mean -0.0317718 +trainer/policy/mean Std 0.904022 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.71729 +trainer/policy/normal/std Std 0.678305 +trainer/policy/normal/std Max 5.68237 +trainer/policy/normal/std Min 0.259668 +trainer/policy/normal/log_std Mean 0.952667 +trainer/policy/normal/log_std Std 0.351402 +trainer/policy/normal/log_std Max 1.73737 +trainer/policy/normal/log_std Min -1.34835 +eval/num steps total 746320 +eval/num paths total 750 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0401003 +eval/Actions Std 0.902672 +eval/Actions Max 0.999995 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56615 +time/logging (s) 0.00403091 +time/sampling batch (s) 0.2802 +time/saving (s) 0.00402342 +time/training (s) 7.40156 +time/epoch (s) 10.256 +time/total (s) 7661.16 +Epoch -251 +---------------------------------- --------------- +2022-05-10 15:18:37.809211 PDT | [0] Epoch -250 finished +---------------------------------- --------------- +epoch -250 +replay_buffer/size 999033 +trainer/num train calls 751000 +trainer/Policy Loss -19.2918 +trainer/Log Pis Mean 26.0062 +trainer/Log Pis Std 14.2047 +trainer/Log Pis Max 83.1411 +trainer/Log Pis Min -9.03268 +trainer/policy/mean Mean -0.0393395 +trainer/policy/mean Std 0.909999 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79551 +trainer/policy/normal/std Std 0.685081 +trainer/policy/normal/std Max 6.02226 +trainer/policy/normal/std Min 0.318451 +trainer/policy/normal/log_std Mean 0.983274 +trainer/policy/normal/log_std Std 0.340497 +trainer/policy/normal/log_std Max 1.79546 +trainer/policy/normal/log_std Min -1.14429 +eval/num steps total 747320 +eval/num paths total 751 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128152 +eval/Actions Std 0.910612 +eval/Actions Max 0.999994 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57119 +time/logging (s) 0.00380471 +time/sampling batch (s) 0.534735 +time/saving (s) 0.00354755 +time/training (s) 6.5813 +time/epoch (s) 9.69457 +time/total (s) 7670.86 +Epoch -250 +---------------------------------- --------------- +2022-05-10 15:18:47.654894 PDT | [0] Epoch -249 finished +---------------------------------- --------------- +epoch -249 +replay_buffer/size 999033 +trainer/num train calls 752000 +trainer/Policy Loss -18.7959 +trainer/Log Pis Mean 24.4523 +trainer/Log Pis Std 13.4817 +trainer/Log Pis Max 64.2896 +trainer/Log Pis Min -12.9393 +trainer/policy/mean Mean -0.0282986 +trainer/policy/mean Std 0.902025 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999968 +trainer/policy/normal/std Mean 2.7439 +trainer/policy/normal/std Std 0.663796 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.287254 +trainer/policy/normal/log_std Mean 0.966171 +trainer/policy/normal/log_std Std 0.334759 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24739 +eval/num steps total 748320 +eval/num paths total 752 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0362966 +eval/Actions Std 0.903947 +eval/Actions Max 0.999999 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52995 +time/logging (s) 0.00374604 +time/sampling batch (s) 0.277937 +time/saving (s) 0.0034444 +time/training (s) 7.0078 +time/epoch (s) 9.82287 +time/total (s) 7680.68 +Epoch -249 +---------------------------------- --------------- +2022-05-10 15:18:57.710806 PDT | [0] Epoch -248 finished +---------------------------------- --------------- +epoch -248 +replay_buffer/size 999033 +trainer/num train calls 753000 +trainer/Policy Loss -19.3867 +trainer/Log Pis Mean 23.5941 +trainer/Log Pis Std 13.1173 +trainer/Log Pis Max 66.4817 +trainer/Log Pis Min -11.0449 +trainer/policy/mean Mean -0.0465807 +trainer/policy/mean Std 0.910496 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79136 +trainer/policy/normal/std Std 0.637527 +trainer/policy/normal/std Max 5.58062 +trainer/policy/normal/std Min 0.30661 +trainer/policy/normal/log_std Mean 0.988955 +trainer/policy/normal/log_std Std 0.309011 +trainer/policy/normal/log_std Max 1.7193 +trainer/policy/normal/log_std Min -1.18218 +eval/num steps total 749320 +eval/num paths total 753 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115419 +eval/Actions Std 0.912277 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64052 +time/logging (s) 0.00391357 +time/sampling batch (s) 0.279804 +time/saving (s) 0.00343364 +time/training (s) 7.10548 +time/epoch (s) 10.0332 +time/total (s) 7690.72 +Epoch -248 +---------------------------------- --------------- +2022-05-10 15:19:07.340608 PDT | [0] Epoch -247 finished +---------------------------------- --------------- +epoch -247 +replay_buffer/size 999033 +trainer/num train calls 754000 +trainer/Policy Loss -19.5545 +trainer/Log Pis Mean 24.6502 +trainer/Log Pis Std 13.4789 +trainer/Log Pis Max 75.1085 +trainer/Log Pis Min -8.46753 +trainer/policy/mean Mean -0.0443956 +trainer/policy/mean Std 0.904953 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.7995 +trainer/policy/normal/std Std 0.682432 +trainer/policy/normal/std Max 5.96062 +trainer/policy/normal/std Min 0.270836 +trainer/policy/normal/log_std Mean 0.985644 +trainer/policy/normal/log_std Std 0.336115 +trainer/policy/normal/log_std Max 1.78517 +trainer/policy/normal/log_std Min -1.30624 +eval/num steps total 750320 +eval/num paths total 754 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.078337 +eval/Actions Std 0.914795 +eval/Actions Max 0.999997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53131 +time/logging (s) 0.00369621 +time/sampling batch (s) 0.281669 +time/saving (s) 0.00346455 +time/training (s) 6.78619 +time/epoch (s) 9.60633 +time/total (s) 7700.33 +Epoch -247 +---------------------------------- --------------- +2022-05-10 15:19:17.961628 PDT | [0] Epoch -246 finished +---------------------------------- --------------- +epoch -246 +replay_buffer/size 999033 +trainer/num train calls 755000 +trainer/Policy Loss -19.0981 +trainer/Log Pis Mean 23.9295 +trainer/Log Pis Std 13.3455 +trainer/Log Pis Max 68.6516 +trainer/Log Pis Min -10.2745 +trainer/policy/mean Mean -0.0401976 +trainer/policy/mean Std 0.90284 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79453 +trainer/policy/normal/std Std 0.661947 +trainer/policy/normal/std Max 6.20924 +trainer/policy/normal/std Min 0.203666 +trainer/policy/normal/log_std Mean 0.988062 +trainer/policy/normal/log_std Std 0.316395 +trainer/policy/normal/log_std Max 1.82604 +trainer/policy/normal/log_std Min -1.59127 +eval/num steps total 751320 +eval/num paths total 755 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.317298 +eval/Actions Std 0.861293 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69348 +time/logging (s) 0.00420894 +time/sampling batch (s) 0.282078 +time/saving (s) 0.0041608 +time/training (s) 7.6141 +time/epoch (s) 10.598 +time/total (s) 7710.93 +Epoch -246 +---------------------------------- --------------- +2022-05-10 15:19:27.364854 PDT | [0] Epoch -245 finished +---------------------------------- --------------- +epoch -245 +replay_buffer/size 999033 +trainer/num train calls 756000 +trainer/Policy Loss -19.8521 +trainer/Log Pis Mean 24.3555 +trainer/Log Pis Std 13.9544 +trainer/Log Pis Max 72.8487 +trainer/Log Pis Min -6.59443 +trainer/policy/mean Mean -0.0315901 +trainer/policy/mean Std 0.906446 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.81861 +trainer/policy/normal/std Std 0.664452 +trainer/policy/normal/std Max 6.01756 +trainer/policy/normal/std Min 0.260854 +trainer/policy/normal/log_std Mean 0.994353 +trainer/policy/normal/log_std Std 0.33291 +trainer/policy/normal/log_std Max 1.79468 +trainer/policy/normal/log_std Min -1.34379 +eval/num steps total 752320 +eval/num paths total 756 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0420165 +eval/Actions Std 0.907841 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49085 +time/logging (s) 0.00379881 +time/sampling batch (s) 0.278712 +time/saving (s) 0.00358721 +time/training (s) 6.60253 +time/epoch (s) 9.37948 +time/total (s) 7720.31 +Epoch -245 +---------------------------------- --------------- +2022-05-10 15:19:37.215183 PDT | [0] Epoch -244 finished +---------------------------------- --------------- +epoch -244 +replay_buffer/size 999033 +trainer/num train calls 757000 +trainer/Policy Loss -19.9125 +trainer/Log Pis Mean 24.9962 +trainer/Log Pis Std 13.0123 +trainer/Log Pis Max 68.5975 +trainer/Log Pis Min -11.4776 +trainer/policy/mean Mean -0.029666 +trainer/policy/mean Std 0.909633 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78615 +trainer/policy/normal/std Std 0.684534 +trainer/policy/normal/std Max 5.72677 +trainer/policy/normal/std Min 0.257837 +trainer/policy/normal/log_std Mean 0.979677 +trainer/policy/normal/log_std Std 0.342336 +trainer/policy/normal/log_std Max 1.74515 +trainer/policy/normal/log_std Min -1.35543 +eval/num steps total 753320 +eval/num paths total 757 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0488642 +eval/Actions Std 0.87609 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62047 +time/logging (s) 0.00387187 +time/sampling batch (s) 0.284015 +time/saving (s) 0.00344089 +time/training (s) 6.91507 +time/epoch (s) 9.82687 +time/total (s) 7730.14 +Epoch -244 +---------------------------------- --------------- +2022-05-10 15:19:46.332128 PDT | [0] Epoch -243 finished +---------------------------------- --------------- +epoch -243 +replay_buffer/size 999033 +trainer/num train calls 758000 +trainer/Policy Loss -19.3832 +trainer/Log Pis Mean 24.7534 +trainer/Log Pis Std 13.4721 +trainer/Log Pis Max 71.1566 +trainer/Log Pis Min -8.36377 +trainer/policy/mean Mean -0.0227796 +trainer/policy/mean Std 0.906335 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80688 +trainer/policy/normal/std Std 0.702271 +trainer/policy/normal/std Max 4.96578 +trainer/policy/normal/std Min 0.238767 +trainer/policy/normal/log_std Mean 0.983195 +trainer/policy/normal/log_std Std 0.362134 +trainer/policy/normal/log_std Max 1.60257 +trainer/policy/normal/log_std Min -1.43227 +eval/num steps total 754320 +eval/num paths total 758 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0961889 +eval/Actions Std 0.90309 +eval/Actions Max 0.999995 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39409 +time/logging (s) 0.00374323 +time/sampling batch (s) 0.283614 +time/saving (s) 0.0034212 +time/training (s) 6.40865 +time/epoch (s) 9.09352 +time/total (s) 7739.24 +Epoch -243 +---------------------------------- --------------- +2022-05-10 15:19:55.846545 PDT | [0] Epoch -242 finished +---------------------------------- --------------- +epoch -242 +replay_buffer/size 999033 +trainer/num train calls 759000 +trainer/Policy Loss -19.8207 +trainer/Log Pis Mean 25.2516 +trainer/Log Pis Std 13.2043 +trainer/Log Pis Max 62.1761 +trainer/Log Pis Min -7.73508 +trainer/policy/mean Mean -0.0212286 +trainer/policy/mean Std 0.910087 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79551 +trainer/policy/normal/std Std 0.660937 +trainer/policy/normal/std Max 5.41363 +trainer/policy/normal/std Min 0.226828 +trainer/policy/normal/log_std Mean 0.98613 +trainer/policy/normal/log_std Std 0.331112 +trainer/policy/normal/log_std Max 1.68892 +trainer/policy/normal/log_std Min -1.48356 +eval/num steps total 755320 +eval/num paths total 759 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0153941 +eval/Actions Std 0.911419 +eval/Actions Max 1 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41137 +time/logging (s) 0.00376329 +time/sampling batch (s) 0.289967 +time/saving (s) 0.00358869 +time/training (s) 6.78204 +time/epoch (s) 9.49073 +time/total (s) 7748.73 +Epoch -242 +---------------------------------- --------------- +2022-05-10 15:20:06.414377 PDT | [0] Epoch -241 finished +---------------------------------- --------------- +epoch -241 +replay_buffer/size 999033 +trainer/num train calls 760000 +trainer/Policy Loss -20.1255 +trainer/Log Pis Mean 24.4338 +trainer/Log Pis Std 12.4143 +trainer/Log Pis Max 62.9908 +trainer/Log Pis Min -11.7562 +trainer/policy/mean Mean -0.0239968 +trainer/policy/mean Std 0.907606 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81573 +trainer/policy/normal/std Std 0.683991 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.223976 +trainer/policy/normal/log_std Mean 0.992143 +trainer/policy/normal/log_std Std 0.33471 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.49622 +eval/num steps total 756320 +eval/num paths total 760 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0347151 +eval/Actions Std 0.905274 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54998 +time/logging (s) 0.00370925 +time/sampling batch (s) 0.536765 +time/saving (s) 0.00351459 +time/training (s) 7.4501 +time/epoch (s) 10.5441 +time/total (s) 7759.28 +Epoch -241 +---------------------------------- --------------- +2022-05-10 15:20:16.168218 PDT | [0] Epoch -240 finished +---------------------------------- --------------- +epoch -240 +replay_buffer/size 999033 +trainer/num train calls 761000 +trainer/Policy Loss -19.3647 +trainer/Log Pis Mean 24.0837 +trainer/Log Pis Std 13.0841 +trainer/Log Pis Max 69.5021 +trainer/Log Pis Min -7.18473 +trainer/policy/mean Mean -0.0399096 +trainer/policy/mean Std 0.905789 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7604 +trainer/policy/normal/std Std 0.664137 +trainer/policy/normal/std Max 5.12581 +trainer/policy/normal/std Min 0.228442 +trainer/policy/normal/log_std Mean 0.970069 +trainer/policy/normal/log_std Std 0.348455 +trainer/policy/normal/log_std Max 1.63429 +trainer/policy/normal/log_std Min -1.47647 +eval/num steps total 757320 +eval/num paths total 761 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0238231 +eval/Actions Std 0.900962 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64526 +time/logging (s) 0.00401951 +time/sampling batch (s) 0.782251 +time/saving (s) 0.00417171 +time/training (s) 6.29517 +time/epoch (s) 9.73087 +time/total (s) 7769.01 +Epoch -240 +---------------------------------- --------------- +2022-05-10 15:20:26.155106 PDT | [0] Epoch -239 finished +---------------------------------- --------------- +epoch -239 +replay_buffer/size 999033 +trainer/num train calls 762000 +trainer/Policy Loss -18.6327 +trainer/Log Pis Mean 24.7098 +trainer/Log Pis Std 13.8945 +trainer/Log Pis Max 73.8542 +trainer/Log Pis Min -11.9117 +trainer/policy/mean Mean -0.0428642 +trainer/policy/mean Std 0.907912 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77559 +trainer/policy/normal/std Std 0.637181 +trainer/policy/normal/std Max 5.77792 +trainer/policy/normal/std Min 0.25592 +trainer/policy/normal/log_std Mean 0.98275 +trainer/policy/normal/log_std Std 0.312565 +trainer/policy/normal/log_std Max 1.75404 +trainer/policy/normal/log_std Min -1.36289 +eval/num steps total 758320 +eval/num paths total 762 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.121763 +eval/Actions Std 0.907507 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76926 +time/logging (s) 0.00384239 +time/sampling batch (s) 0.281149 +time/saving (s) 0.00353241 +time/training (s) 6.90538 +time/epoch (s) 9.96317 +time/total (s) 7778.98 +Epoch -239 +---------------------------------- --------------- +2022-05-10 15:20:35.324339 PDT | [0] Epoch -238 finished +---------------------------------- --------------- +epoch -238 +replay_buffer/size 999033 +trainer/num train calls 763000 +trainer/Policy Loss -20.5718 +trainer/Log Pis Mean 24.7694 +trainer/Log Pis Std 13.4975 +trainer/Log Pis Max 65.1083 +trainer/Log Pis Min -7.23914 +trainer/policy/mean Mean -0.034241 +trainer/policy/mean Std 0.907215 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80012 +trainer/policy/normal/std Std 0.668938 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.259823 +trainer/policy/normal/log_std Mean 0.987605 +trainer/policy/normal/log_std Std 0.329522 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.34775 +eval/num steps total 759320 +eval/num paths total 763 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.197853 +eval/Actions Std 0.954016 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71 +time/logging (s) 0.00377516 +time/sampling batch (s) 0.277416 +time/saving (s) 0.00343236 +time/training (s) 6.15158 +time/epoch (s) 9.14621 +time/total (s) 7788.13 +Epoch -238 +---------------------------------- --------------- +2022-05-10 15:20:45.269138 PDT | [0] Epoch -237 finished +---------------------------------- --------------- +epoch -237 +replay_buffer/size 999033 +trainer/num train calls 764000 +trainer/Policy Loss -19.074 +trainer/Log Pis Mean 23.9003 +trainer/Log Pis Std 13.5753 +trainer/Log Pis Max 69.2517 +trainer/Log Pis Min -10.2746 +trainer/policy/mean Mean -0.0284936 +trainer/policy/mean Std 0.904794 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78096 +trainer/policy/normal/std Std 0.67531 +trainer/policy/normal/std Max 6.10584 +trainer/policy/normal/std Min 0.26842 +trainer/policy/normal/log_std Mean 0.979761 +trainer/policy/normal/log_std Std 0.332562 +trainer/policy/normal/log_std Max 1.80925 +trainer/policy/normal/log_std Min -1.3152 +eval/num steps total 760320 +eval/num paths total 764 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.115414 +eval/Actions Std 0.900579 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73799 +time/logging (s) 0.00385137 +time/sampling batch (s) 0.290788 +time/saving (s) 0.00356563 +time/training (s) 6.88518 +time/epoch (s) 9.92137 +time/total (s) 7798.05 +Epoch -237 +---------------------------------- --------------- +2022-05-10 15:20:54.708877 PDT | [0] Epoch -236 finished +---------------------------------- --------------- +epoch -236 +replay_buffer/size 999033 +trainer/num train calls 765000 +trainer/Policy Loss -19.5692 +trainer/Log Pis Mean 24.5113 +trainer/Log Pis Std 12.8382 +trainer/Log Pis Max 64.3913 +trainer/Log Pis Min -9.55293 +trainer/policy/mean Mean -0.0434572 +trainer/policy/mean Std 0.898247 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.7466 +trainer/policy/normal/std Std 0.663582 +trainer/policy/normal/std Max 5.48461 +trainer/policy/normal/std Min 0.311221 +trainer/policy/normal/log_std Mean 0.968176 +trainer/policy/normal/log_std Std 0.327261 +trainer/policy/normal/log_std Max 1.70195 +trainer/policy/normal/log_std Min -1.16725 +eval/num steps total 761320 +eval/num paths total 765 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.441089 +eval/Actions Std 0.844453 +eval/Actions Max 0.999984 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77071 +time/logging (s) 0.00378543 +time/sampling batch (s) 0.291379 +time/saving (s) 0.00352174 +time/training (s) 6.34625 +time/epoch (s) 9.41565 +time/total (s) 7807.47 +Epoch -236 +---------------------------------- --------------- +2022-05-10 15:21:04.152461 PDT | [0] Epoch -235 finished +---------------------------------- --------------- +epoch -235 +replay_buffer/size 999033 +trainer/num train calls 766000 +trainer/Policy Loss -20.1742 +trainer/Log Pis Mean 25.054 +trainer/Log Pis Std 13.7861 +trainer/Log Pis Max 70.1174 +trainer/Log Pis Min -7.87439 +trainer/policy/mean Mean -0.0384151 +trainer/policy/mean Std 0.904585 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.8063 +trainer/policy/normal/std Std 0.647204 +trainer/policy/normal/std Max 5.5844 +trainer/policy/normal/std Min 0.26839 +trainer/policy/normal/log_std Mean 0.992164 +trainer/policy/normal/log_std Std 0.320685 +trainer/policy/normal/log_std Max 1.71998 +trainer/policy/normal/log_std Min -1.31531 +eval/num steps total 762320 +eval/num paths total 766 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.145923 +eval/Actions Std 0.87823 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64545 +time/logging (s) 0.00386856 +time/sampling batch (s) 0.542986 +time/saving (s) 0.00353395 +time/training (s) 6.22386 +time/epoch (s) 9.4197 +time/total (s) 7816.89 +Epoch -235 +---------------------------------- --------------- +2022-05-10 15:21:14.405119 PDT | [0] Epoch -234 finished +---------------------------------- --------------- +epoch -234 +replay_buffer/size 999033 +trainer/num train calls 767000 +trainer/Policy Loss -19.3458 +trainer/Log Pis Mean 24.06 +trainer/Log Pis Std 13.6992 +trainer/Log Pis Max 63.6581 +trainer/Log Pis Min -6.65592 +trainer/policy/mean Mean -0.0339419 +trainer/policy/mean Std 0.905735 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85272 +trainer/policy/normal/std Std 0.690649 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.325143 +trainer/policy/normal/log_std Mean 1.00573 +trainer/policy/normal/log_std Std 0.330081 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.12349 +eval/num steps total 763320 +eval/num paths total 767 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0159015 +eval/Actions Std 0.90505 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70384 +time/logging (s) 0.00414727 +time/sampling batch (s) 0.287039 +time/saving (s) 0.00410599 +time/training (s) 7.23007 +time/epoch (s) 10.2292 +time/total (s) 7827.12 +Epoch -234 +---------------------------------- --------------- +2022-05-10 15:21:24.747872 PDT | [0] Epoch -233 finished +---------------------------------- --------------- +epoch -233 +replay_buffer/size 999033 +trainer/num train calls 768000 +trainer/Policy Loss -19.8404 +trainer/Log Pis Mean 24.7716 +trainer/Log Pis Std 13.0838 +trainer/Log Pis Max 70.1803 +trainer/Log Pis Min -3.57236 +trainer/policy/mean Mean -0.0382353 +trainer/policy/mean Std 0.906802 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84583 +trainer/policy/normal/std Std 0.662114 +trainer/policy/normal/std Max 5.64688 +trainer/policy/normal/std Min 0.294353 +trainer/policy/normal/log_std Mean 1.00664 +trainer/policy/normal/log_std Std 0.316791 +trainer/policy/normal/log_std Max 1.7311 +trainer/policy/normal/log_std Min -1.22298 +eval/num steps total 764320 +eval/num paths total 768 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.130866 +eval/Actions Std 0.872372 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60428 +time/logging (s) 0.00380235 +time/sampling batch (s) 0.53186 +time/saving (s) 0.00353942 +time/training (s) 7.17557 +time/epoch (s) 10.3191 +time/total (s) 7837.45 +Epoch -233 +---------------------------------- --------------- +2022-05-10 15:21:34.957872 PDT | [0] Epoch -232 finished +---------------------------------- --------------- +epoch -232 +replay_buffer/size 999033 +trainer/num train calls 769000 +trainer/Policy Loss -20.4977 +trainer/Log Pis Mean 23.6344 +trainer/Log Pis Std 13.0018 +trainer/Log Pis Max 65.9608 +trainer/Log Pis Min -5.18621 +trainer/policy/mean Mean -0.0433594 +trainer/policy/mean Std 0.907269 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80172 +trainer/policy/normal/std Std 0.675772 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.298848 +trainer/policy/normal/log_std Mean 0.987339 +trainer/policy/normal/log_std Std 0.333998 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.20782 +eval/num steps total 765320 +eval/num paths total 769 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0466901 +eval/Actions Std 0.91237 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50716 +time/logging (s) 0.00379522 +time/sampling batch (s) 0.283919 +time/saving (s) 0.00346067 +time/training (s) 7.38827 +time/epoch (s) 10.1866 +time/total (s) 7847.64 +Epoch -232 +---------------------------------- --------------- +2022-05-10 15:21:45.312380 PDT | [0] Epoch -231 finished +---------------------------------- --------------- +epoch -231 +replay_buffer/size 999033 +trainer/num train calls 770000 +trainer/Policy Loss -19.2906 +trainer/Log Pis Mean 23.0532 +trainer/Log Pis Std 13.3397 +trainer/Log Pis Max 64.3659 +trainer/Log Pis Min -7.50003 +trainer/policy/mean Mean -0.033647 +trainer/policy/mean Std 0.903198 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.68128 +trainer/policy/normal/std Std 0.641887 +trainer/policy/normal/std Max 5.29628 +trainer/policy/normal/std Min 0.225695 +trainer/policy/normal/log_std Mean 0.94458 +trainer/policy/normal/log_std Std 0.326341 +trainer/policy/normal/log_std Max 1.667 +trainer/policy/normal/log_std Min -1.48857 +eval/num steps total 766320 +eval/num paths total 770 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0307752 +eval/Actions Std 0.918477 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6148 +time/logging (s) 0.00384178 +time/sampling batch (s) 0.285143 +time/saving (s) 0.00341304 +time/training (s) 7.4239 +time/epoch (s) 10.3311 +time/total (s) 7857.97 +Epoch -231 +---------------------------------- --------------- +2022-05-10 15:21:55.323544 PDT | [0] Epoch -230 finished +---------------------------------- --------------- +epoch -230 +replay_buffer/size 999033 +trainer/num train calls 771000 +trainer/Policy Loss -19.8224 +trainer/Log Pis Mean 24.8292 +trainer/Log Pis Std 13.5551 +trainer/Log Pis Max 86.6501 +trainer/Log Pis Min -11.0312 +trainer/policy/mean Mean -0.0313152 +trainer/policy/mean Std 0.912613 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86867 +trainer/policy/normal/std Std 0.639048 +trainer/policy/normal/std Max 6.73716 +trainer/policy/normal/std Min 0.34732 +trainer/policy/normal/log_std Mean 1.01819 +trainer/policy/normal/log_std Std 0.301189 +trainer/policy/normal/log_std Max 1.90764 +trainer/policy/normal/log_std Min -1.05751 +eval/num steps total 767320 +eval/num paths total 771 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123436 +eval/Actions Std 0.887146 +eval/Actions Max 0.999987 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63464 +time/logging (s) 0.00368123 +time/sampling batch (s) 0.283045 +time/saving (s) 0.00343261 +time/training (s) 7.06285 +time/epoch (s) 9.98765 +time/total (s) 7867.96 +Epoch -230 +---------------------------------- --------------- +2022-05-10 15:22:05.964272 PDT | [0] Epoch -229 finished +---------------------------------- --------------- +epoch -229 +replay_buffer/size 999033 +trainer/num train calls 772000 +trainer/Policy Loss -20.5253 +trainer/Log Pis Mean 24.8166 +trainer/Log Pis Std 13.5821 +trainer/Log Pis Max 65.7145 +trainer/Log Pis Min -6.87618 +trainer/policy/mean Mean -0.0537183 +trainer/policy/mean Std 0.902714 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.8584 +trainer/policy/normal/std Std 0.694302 +trainer/policy/normal/std Max 6.07812 +trainer/policy/normal/std Min 0.300051 +trainer/policy/normal/log_std Mean 1.00658 +trainer/policy/normal/log_std Std 0.337272 +trainer/policy/normal/log_std Max 1.8047 +trainer/policy/normal/log_std Min -1.2038 +eval/num steps total 768320 +eval/num paths total 772 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.110525 +eval/Actions Std 0.922133 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56775 +time/logging (s) 0.00370867 +time/sampling batch (s) 0.281459 +time/saving (s) 0.00344437 +time/training (s) 7.76129 +time/epoch (s) 10.6177 +time/total (s) 7878.58 +Epoch -229 +---------------------------------- --------------- +2022-05-10 15:22:15.830951 PDT | [0] Epoch -228 finished +---------------------------------- --------------- +epoch -228 +replay_buffer/size 999033 +trainer/num train calls 773000 +trainer/Policy Loss -19.7895 +trainer/Log Pis Mean 24.3422 +trainer/Log Pis Std 13.606 +trainer/Log Pis Max 67.7188 +trainer/Log Pis Min -3.64848 +trainer/policy/mean Mean -0.0390556 +trainer/policy/mean Std 0.903506 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.80352 +trainer/policy/normal/std Std 0.668111 +trainer/policy/normal/std Max 5.40529 +trainer/policy/normal/std Min 0.287019 +trainer/policy/normal/log_std Mean 0.989749 +trainer/policy/normal/log_std Std 0.324052 +trainer/policy/normal/log_std Max 1.68738 +trainer/policy/normal/log_std Min -1.24821 +eval/num steps total 769320 +eval/num paths total 773 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101518 +eval/Actions Std 0.891799 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58843 +time/logging (s) 0.00426707 +time/sampling batch (s) 0.288347 +time/saving (s) 0.00412833 +time/training (s) 6.95846 +time/epoch (s) 9.84364 +time/total (s) 7888.43 +Epoch -228 +---------------------------------- --------------- +2022-05-10 15:22:25.957494 PDT | [0] Epoch -227 finished +---------------------------------- --------------- +epoch -227 +replay_buffer/size 999033 +trainer/num train calls 774000 +trainer/Policy Loss -19.9143 +trainer/Log Pis Mean 24.0021 +trainer/Log Pis Std 13.6418 +trainer/Log Pis Max 69.5516 +trainer/Log Pis Min -6.59963 +trainer/policy/mean Mean -0.040927 +trainer/policy/mean Std 0.908143 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76862 +trainer/policy/normal/std Std 0.67894 +trainer/policy/normal/std Max 5.14762 +trainer/policy/normal/std Min 0.251996 +trainer/policy/normal/log_std Mean 0.972412 +trainer/policy/normal/log_std Std 0.347883 +trainer/policy/normal/log_std Max 1.63854 +trainer/policy/normal/log_std Min -1.37834 +eval/num steps total 770320 +eval/num paths total 774 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0137239 +eval/Actions Std 0.959218 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59453 +time/logging (s) 0.00376441 +time/sampling batch (s) 0.283021 +time/saving (s) 0.00371599 +time/training (s) 7.21676 +time/epoch (s) 10.1018 +time/total (s) 7898.53 +Epoch -227 +---------------------------------- --------------- +2022-05-10 15:22:35.795568 PDT | [0] Epoch -226 finished +---------------------------------- --------------- +epoch -226 +replay_buffer/size 999033 +trainer/num train calls 775000 +trainer/Policy Loss -19.3784 +trainer/Log Pis Mean 24.9184 +trainer/Log Pis Std 13.422 +trainer/Log Pis Max 74.2832 +trainer/Log Pis Min -7.52252 +trainer/policy/mean Mean -0.0468067 +trainer/policy/mean Std 0.901517 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78578 +trainer/policy/normal/std Std 0.682384 +trainer/policy/normal/std Max 6.2843 +trainer/policy/normal/std Min 0.310462 +trainer/policy/normal/log_std Mean 0.980204 +trainer/policy/normal/log_std Std 0.338029 +trainer/policy/normal/log_std Max 1.83805 +trainer/policy/normal/log_std Min -1.16969 +eval/num steps total 771320 +eval/num paths total 775 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.406758 +eval/Actions Std 0.831588 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55694 +time/logging (s) 0.00373328 +time/sampling batch (s) 0.278354 +time/saving (s) 0.00347655 +time/training (s) 6.97233 +time/epoch (s) 9.81484 +time/total (s) 7908.35 +Epoch -226 +---------------------------------- --------------- +2022-05-10 15:22:46.434560 PDT | [0] Epoch -225 finished +---------------------------------- --------------- +epoch -225 +replay_buffer/size 999033 +trainer/num train calls 776000 +trainer/Policy Loss -19.7259 +trainer/Log Pis Mean 24.1241 +trainer/Log Pis Std 13.8557 +trainer/Log Pis Max 69.8868 +trainer/Log Pis Min -6.54014 +trainer/policy/mean Mean -0.0484466 +trainer/policy/mean Std 0.9027 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.85522 +trainer/policy/normal/std Std 0.689268 +trainer/policy/normal/std Max 6.28368 +trainer/policy/normal/std Min 0.278822 +trainer/policy/normal/log_std Mean 1.00655 +trainer/policy/normal/log_std Std 0.33106 +trainer/policy/normal/log_std Max 1.83796 +trainer/policy/normal/log_std Min -1.27718 +eval/num steps total 772320 +eval/num paths total 776 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.132651 +eval/Actions Std 0.863991 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64866 +time/logging (s) 0.00377368 +time/sampling batch (s) 0.280173 +time/saving (s) 0.00343003 +time/training (s) 7.67982 +time/epoch (s) 10.6159 +time/total (s) 7918.97 +Epoch -225 +---------------------------------- --------------- +2022-05-10 15:22:56.323111 PDT | [0] Epoch -224 finished +---------------------------------- --------------- +epoch -224 +replay_buffer/size 999033 +trainer/num train calls 777000 +trainer/Policy Loss -19.7402 +trainer/Log Pis Mean 23.8641 +trainer/Log Pis Std 12.8713 +trainer/Log Pis Max 66.9829 +trainer/Log Pis Min -9.29798 +trainer/policy/mean Mean -0.0271194 +trainer/policy/mean Std 0.900997 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.84942 +trainer/policy/normal/std Std 0.715516 +trainer/policy/normal/std Max 6.76428 +trainer/policy/normal/std Min 0.244334 +trainer/policy/normal/log_std Mean 1.00125 +trainer/policy/normal/log_std Std 0.345049 +trainer/policy/normal/log_std Max 1.91166 +trainer/policy/normal/log_std Min -1.40922 +eval/num steps total 773320 +eval/num paths total 777 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100709 +eval/Actions Std 0.914119 +eval/Actions Max 0.999995 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55154 +time/logging (s) 0.00385691 +time/sampling batch (s) 0.279003 +time/saving (s) 0.00345571 +time/training (s) 7.02753 +time/epoch (s) 9.86539 +time/total (s) 7928.84 +Epoch -224 +---------------------------------- --------------- +2022-05-10 15:23:06.873188 PDT | [0] Epoch -223 finished +---------------------------------- --------------- +epoch -223 +replay_buffer/size 999033 +trainer/num train calls 778000 +trainer/Policy Loss -20.6694 +trainer/Log Pis Mean 24.9072 +trainer/Log Pis Std 13.826 +trainer/Log Pis Max 74.5725 +trainer/Log Pis Min -9.12017 +trainer/policy/mean Mean -0.034087 +trainer/policy/mean Std 0.907268 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.8309 +trainer/policy/normal/std Std 0.657109 +trainer/policy/normal/std Max 5.40619 +trainer/policy/normal/std Min 0.263233 +trainer/policy/normal/log_std Mean 1.0001 +trainer/policy/normal/log_std Std 0.325648 +trainer/policy/normal/log_std Max 1.68755 +trainer/policy/normal/log_std Min -1.33471 +eval/num steps total 774320 +eval/num paths total 778 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.052174 +eval/Actions Std 0.913547 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49621 +time/logging (s) 0.00382934 +time/sampling batch (s) 0.528982 +time/saving (s) 0.00349904 +time/training (s) 7.49443 +time/epoch (s) 10.527 +time/total (s) 7939.37 +Epoch -223 +---------------------------------- --------------- +2022-05-10 15:23:16.372952 PDT | [0] Epoch -222 finished +---------------------------------- --------------- +epoch -222 +replay_buffer/size 999033 +trainer/num train calls 779000 +trainer/Policy Loss -19.1344 +trainer/Log Pis Mean 23.893 +trainer/Log Pis Std 13.6942 +trainer/Log Pis Max 74.1765 +trainer/Log Pis Min -7.27086 +trainer/policy/mean Mean -0.0180282 +trainer/policy/mean Std 0.899969 +trainer/policy/mean Max 1 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.77923 +trainer/policy/normal/std Std 0.695122 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.230366 +trainer/policy/normal/log_std Mean 0.976119 +trainer/policy/normal/log_std Std 0.346291 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.46809 +eval/num steps total 775320 +eval/num paths total 779 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.239615 +eval/Actions Std 0.82885 +eval/Actions Max 0.999975 +eval/Actions Min -0.999967 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50526 +time/logging (s) 0.00418985 +time/sampling batch (s) 0.281825 +time/saving (s) 0.00404556 +time/training (s) 6.68121 +time/epoch (s) 9.47653 +time/total (s) 7948.85 +Epoch -222 +---------------------------------- --------------- +2022-05-10 15:23:25.665778 PDT | [0] Epoch -221 finished +---------------------------------- --------------- +epoch -221 +replay_buffer/size 999033 +trainer/num train calls 780000 +trainer/Policy Loss -20.6479 +trainer/Log Pis Mean 24.383 +trainer/Log Pis Std 12.7195 +trainer/Log Pis Max 66.8462 +trainer/Log Pis Min -9.22117 +trainer/policy/mean Mean -0.0518516 +trainer/policy/mean Std 0.907888 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87383 +trainer/policy/normal/std Std 0.692566 +trainer/policy/normal/std Max 6.01609 +trainer/policy/normal/std Min 0.257333 +trainer/policy/normal/log_std Mean 1.01254 +trainer/policy/normal/log_std Std 0.334475 +trainer/policy/normal/log_std Max 1.79444 +trainer/policy/normal/log_std Min -1.35739 +eval/num steps total 776320 +eval/num paths total 780 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.125298 +eval/Actions Std 0.87978 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61618 +time/logging (s) 0.00387986 +time/sampling batch (s) 0.278783 +time/saving (s) 0.00368183 +time/training (s) 6.36642 +time/epoch (s) 9.26894 +time/total (s) 7958.12 +Epoch -221 +---------------------------------- --------------- +2022-05-10 15:23:34.651632 PDT | [0] Epoch -220 finished +---------------------------------- --------------- +epoch -220 +replay_buffer/size 999033 +trainer/num train calls 781000 +trainer/Policy Loss -19.098 +trainer/Log Pis Mean 25.1303 +trainer/Log Pis Std 13.7324 +trainer/Log Pis Max 69.6479 +trainer/Log Pis Min -10.0482 +trainer/policy/mean Mean -0.0423841 +trainer/policy/mean Std 0.909778 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80134 +trainer/policy/normal/std Std 0.654662 +trainer/policy/normal/std Max 5.80253 +trainer/policy/normal/std Min 0.289948 +trainer/policy/normal/log_std Mean 0.989605 +trainer/policy/normal/log_std Std 0.324524 +trainer/policy/normal/log_std Max 1.75829 +trainer/policy/normal/log_std Min -1.23805 +eval/num steps total 777320 +eval/num paths total 781 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.489602 +eval/Actions Std 0.848672 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54821 +time/logging (s) 0.00376356 +time/sampling batch (s) 0.279019 +time/saving (s) 0.00346766 +time/training (s) 6.12814 +time/epoch (s) 8.9626 +time/total (s) 7967.08 +Epoch -220 +---------------------------------- --------------- +2022-05-10 15:23:45.060095 PDT | [0] Epoch -219 finished +---------------------------------- --------------- +epoch -219 +replay_buffer/size 999033 +trainer/num train calls 782000 +trainer/Policy Loss -21.1776 +trainer/Log Pis Mean 24.0978 +trainer/Log Pis Std 13.0305 +trainer/Log Pis Max 68.8364 +trainer/Log Pis Min -8.19268 +trainer/policy/mean Mean -0.0471956 +trainer/policy/mean Std 0.904274 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87972 +trainer/policy/normal/std Std 0.684874 +trainer/policy/normal/std Max 6.6604 +trainer/policy/normal/std Min 0.256356 +trainer/policy/normal/log_std Mean 1.01614 +trainer/policy/normal/log_std Std 0.328192 +trainer/policy/normal/log_std Max 1.89618 +trainer/policy/normal/log_std Min -1.36119 +eval/num steps total 778320 +eval/num paths total 782 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.120894 +eval/Actions Std 0.891034 +eval/Actions Max 0.999983 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77118 +time/logging (s) 0.00366621 +time/sampling batch (s) 0.278312 +time/saving (s) 0.0034532 +time/training (s) 7.32875 +time/epoch (s) 10.3854 +time/total (s) 7977.47 +Epoch -219 +---------------------------------- --------------- +2022-05-10 15:23:54.935713 PDT | [0] Epoch -218 finished +---------------------------------- --------------- +epoch -218 +replay_buffer/size 999033 +trainer/num train calls 783000 +trainer/Policy Loss -19.8081 +trainer/Log Pis Mean 24.4605 +trainer/Log Pis Std 13.3448 +trainer/Log Pis Max 77.0329 +trainer/Log Pis Min -6.57467 +trainer/policy/mean Mean -0.0442876 +trainer/policy/mean Std 0.909427 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7886 +trainer/policy/normal/std Std 0.661204 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.287234 +trainer/policy/normal/log_std Mean 0.986045 +trainer/policy/normal/log_std Std 0.316491 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.24746 +eval/num steps total 779320 +eval/num paths total 783 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0282423 +eval/Actions Std 0.905163 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54223 +time/logging (s) 0.0036746 +time/sampling batch (s) 0.284193 +time/saving (s) 0.00348188 +time/training (s) 7.01883 +time/epoch (s) 9.85241 +time/total (s) 7987.33 +Epoch -218 +---------------------------------- --------------- +2022-05-10 15:24:04.712259 PDT | [0] Epoch -217 finished +---------------------------------- --------------- +epoch -217 +replay_buffer/size 999033 +trainer/num train calls 784000 +trainer/Policy Loss -19.3115 +trainer/Log Pis Mean 23.5947 +trainer/Log Pis Std 13.0624 +trainer/Log Pis Max 64.9521 +trainer/Log Pis Min -7.74986 +trainer/policy/mean Mean -0.0379706 +trainer/policy/mean Std 0.907435 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78647 +trainer/policy/normal/std Std 0.665178 +trainer/policy/normal/std Max 6.40502 +trainer/policy/normal/std Min 0.269742 +trainer/policy/normal/log_std Mean 0.98395 +trainer/policy/normal/log_std Std 0.321307 +trainer/policy/normal/log_std Max 1.85708 +trainer/policy/normal/log_std Min -1.31029 +eval/num steps total 780320 +eval/num paths total 784 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106566 +eval/Actions Std 0.917516 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34166 +time/logging (s) 0.00367858 +time/sampling batch (s) 0.279805 +time/saving (s) 0.00346273 +time/training (s) 7.12464 +time/epoch (s) 9.75324 +time/total (s) 7997.08 +Epoch -217 +---------------------------------- --------------- +2022-05-10 15:24:14.259130 PDT | [0] Epoch -216 finished +---------------------------------- --------------- +epoch -216 +replay_buffer/size 999033 +trainer/num train calls 785000 +trainer/Policy Loss -19.4184 +trainer/Log Pis Mean 24.7907 +trainer/Log Pis Std 12.5647 +trainer/Log Pis Max 61.1678 +trainer/Log Pis Min -1.89289 +trainer/policy/mean Mean -0.0430298 +trainer/policy/mean Std 0.905926 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.85242 +trainer/policy/normal/std Std 0.6919 +trainer/policy/normal/std Max 5.57519 +trainer/policy/normal/std Min 0.254803 +trainer/policy/normal/log_std Mean 1.00284 +trainer/policy/normal/log_std Std 0.347496 +trainer/policy/normal/log_std Max 1.71833 +trainer/policy/normal/log_std Min -1.36726 +eval/num steps total 781320 +eval/num paths total 785 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0173865 +eval/Actions Std 0.911144 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45548 +time/logging (s) 0.00430328 +time/sampling batch (s) 0.280618 +time/saving (s) 0.00417494 +time/training (s) 6.77952 +time/epoch (s) 9.5241 +time/total (s) 8006.61 +Epoch -216 +---------------------------------- --------------- +2022-05-10 15:24:23.757874 PDT | [0] Epoch -215 finished +---------------------------------- --------------- +epoch -215 +replay_buffer/size 999033 +trainer/num train calls 786000 +trainer/Policy Loss -19.4174 +trainer/Log Pis Mean 24.2762 +trainer/Log Pis Std 13.0352 +trainer/Log Pis Max 69.3897 +trainer/Log Pis Min -10.5318 +trainer/policy/mean Mean -0.0442601 +trainer/policy/mean Std 0.904692 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80877 +trainer/policy/normal/std Std 0.671515 +trainer/policy/normal/std Max 6.01591 +trainer/policy/normal/std Min 0.307851 +trainer/policy/normal/log_std Mean 0.990703 +trainer/policy/normal/log_std Std 0.329669 +trainer/policy/normal/log_std Max 1.79441 +trainer/policy/normal/log_std Min -1.17814 +eval/num steps total 782320 +eval/num paths total 786 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.249156 +eval/Actions Std 0.92435 +eval/Actions Max 0.999985 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.85461 +time/logging (s) 0.00376272 +time/sampling batch (s) 0.289997 +time/saving (s) 0.00364832 +time/training (s) 6.32162 +time/epoch (s) 9.47363 +time/total (s) 8016.09 +Epoch -215 +---------------------------------- --------------- +2022-05-10 15:24:33.259878 PDT | [0] Epoch -214 finished +---------------------------------- --------------- +epoch -214 +replay_buffer/size 999033 +trainer/num train calls 787000 +trainer/Policy Loss -20.4295 +trainer/Log Pis Mean 24.8499 +trainer/Log Pis Std 13.7184 +trainer/Log Pis Max 71.1564 +trainer/Log Pis Min -6.28226 +trainer/policy/mean Mean -0.0373111 +trainer/policy/mean Std 0.907167 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80962 +trainer/policy/normal/std Std 0.655757 +trainer/policy/normal/std Max 7.07995 +trainer/policy/normal/std Min 0.284445 +trainer/policy/normal/log_std Mean 0.992787 +trainer/policy/normal/log_std Std 0.323316 +trainer/policy/normal/log_std Max 1.95727 +trainer/policy/normal/log_std Min -1.25722 +eval/num steps total 783320 +eval/num paths total 787 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0385719 +eval/Actions Std 0.931953 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60982 +time/logging (s) 0.00382536 +time/sampling batch (s) 0.281005 +time/saving (s) 0.00343263 +time/training (s) 6.58069 +time/epoch (s) 9.47877 +time/total (s) 8025.57 +Epoch -214 +---------------------------------- --------------- +2022-05-10 15:24:43.592426 PDT | [0] Epoch -213 finished +---------------------------------- --------------- +epoch -213 +replay_buffer/size 999033 +trainer/num train calls 788000 +trainer/Policy Loss -19.3991 +trainer/Log Pis Mean 24.7283 +trainer/Log Pis Std 13.0336 +trainer/Log Pis Max 57.6574 +trainer/Log Pis Min -6.26828 +trainer/policy/mean Mean -0.0238326 +trainer/policy/mean Std 0.904836 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79069 +trainer/policy/normal/std Std 0.663338 +trainer/policy/normal/std Max 5.55432 +trainer/policy/normal/std Min 0.283436 +trainer/policy/normal/log_std Mean 0.98305 +trainer/policy/normal/log_std Std 0.337889 +trainer/policy/normal/log_std Max 1.71458 +trainer/policy/normal/log_std Min -1.26077 +eval/num steps total 784320 +eval/num paths total 788 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.108095 +eval/Actions Std 0.894773 +eval/Actions Max 0.999992 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61781 +time/logging (s) 0.00382525 +time/sampling batch (s) 0.284216 +time/saving (s) 0.00346828 +time/training (s) 7.3998 +time/epoch (s) 10.3091 +time/total (s) 8035.88 +Epoch -213 +---------------------------------- --------------- +2022-05-10 15:24:52.986284 PDT | [0] Epoch -212 finished +---------------------------------- --------------- +epoch -212 +replay_buffer/size 999033 +trainer/num train calls 789000 +trainer/Policy Loss -19.8515 +trainer/Log Pis Mean 24.5713 +trainer/Log Pis Std 13.7665 +trainer/Log Pis Max 70.9966 +trainer/Log Pis Min -8.34118 +trainer/policy/mean Mean -0.0453172 +trainer/policy/mean Std 0.902073 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85924 +trainer/policy/normal/std Std 0.651049 +trainer/policy/normal/std Max 5.69835 +trainer/policy/normal/std Min 0.208648 +trainer/policy/normal/log_std Mean 1.01328 +trainer/policy/normal/log_std Std 0.308802 +trainer/policy/normal/log_std Max 1.74018 +trainer/policy/normal/log_std Min -1.5671 +eval/num steps total 785320 +eval/num paths total 789 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.222673 +eval/Actions Std 0.951141 +eval/Actions Max 0.999995 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65869 +time/logging (s) 0.00371721 +time/sampling batch (s) 0.286635 +time/saving (s) 0.00346854 +time/training (s) 6.41751 +time/epoch (s) 9.37002 +time/total (s) 8045.26 +Epoch -212 +---------------------------------- --------------- +2022-05-10 15:25:02.963942 PDT | [0] Epoch -211 finished +---------------------------------- --------------- +epoch -211 +replay_buffer/size 999033 +trainer/num train calls 790000 +trainer/Policy Loss -20.0067 +trainer/Log Pis Mean 24.9101 +trainer/Log Pis Std 13.1465 +trainer/Log Pis Max 70.3762 +trainer/Log Pis Min -8.86828 +trainer/policy/mean Mean -0.0535422 +trainer/policy/mean Std 0.907034 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8013 +trainer/policy/normal/std Std 0.660292 +trainer/policy/normal/std Max 6.2166 +trainer/policy/normal/std Min 0.260718 +trainer/policy/normal/log_std Mean 0.989571 +trainer/policy/normal/log_std Std 0.322101 +trainer/policy/normal/log_std Max 1.82722 +trainer/policy/normal/log_std Min -1.34432 +eval/num steps total 786320 +eval/num paths total 790 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00064551 +eval/Actions Std 0.903439 +eval/Actions Max 0.999995 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47502 +time/logging (s) 0.00369794 +time/sampling batch (s) 0.283736 +time/saving (s) 0.00344443 +time/training (s) 7.18843 +time/epoch (s) 9.95433 +time/total (s) 8055.21 +Epoch -211 +---------------------------------- --------------- +2022-05-10 15:25:13.450471 PDT | [0] Epoch -210 finished +---------------------------------- --------------- +epoch -210 +replay_buffer/size 999033 +trainer/num train calls 791000 +trainer/Policy Loss -18.9923 +trainer/Log Pis Mean 24.9509 +trainer/Log Pis Std 13.7177 +trainer/Log Pis Max 73.5476 +trainer/Log Pis Min -8.9861 +trainer/policy/mean Mean -0.0360339 +trainer/policy/mean Std 0.906638 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.84058 +trainer/policy/normal/std Std 0.681287 +trainer/policy/normal/std Max 5.88449 +trainer/policy/normal/std Min 0.314228 +trainer/policy/normal/log_std Mean 1.00128 +trainer/policy/normal/log_std Std 0.333292 +trainer/policy/normal/log_std Max 1.77232 +trainer/policy/normal/log_std Min -1.15764 +eval/num steps total 787320 +eval/num paths total 791 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0643557 +eval/Actions Std 0.904611 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58168 +time/logging (s) 0.00415083 +time/sampling batch (s) 0.281252 +time/saving (s) 0.00411217 +time/training (s) 7.59235 +time/epoch (s) 10.4636 +time/total (s) 8065.68 +Epoch -210 +---------------------------------- --------------- +2022-05-10 15:25:24.585328 PDT | [0] Epoch -209 finished +---------------------------------- --------------- +epoch -209 +replay_buffer/size 999033 +trainer/num train calls 792000 +trainer/Policy Loss -19.7789 +trainer/Log Pis Mean 24.3704 +trainer/Log Pis Std 14.1021 +trainer/Log Pis Max 76.8108 +trainer/Log Pis Min -10.6772 +trainer/policy/mean Mean -0.0391809 +trainer/policy/mean Std 0.901673 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81116 +trainer/policy/normal/std Std 0.69849 +trainer/policy/normal/std Max 6.85069 +trainer/policy/normal/std Min 0.280754 +trainer/policy/normal/log_std Mean 0.98751 +trainer/policy/normal/log_std Std 0.346404 +trainer/policy/normal/log_std Max 1.92435 +trainer/policy/normal/log_std Min -1.27028 +eval/num steps total 788320 +eval/num paths total 792 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.379774 +eval/Actions Std 0.684997 +eval/Actions Max 0.999996 +eval/Actions Min -0.999965 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54881 +time/logging (s) 0.00387998 +time/sampling batch (s) 0.285053 +time/saving (s) 0.0036515 +time/training (s) 8.26932 +time/epoch (s) 11.1107 +time/total (s) 8076.79 +Epoch -209 +---------------------------------- --------------- +2022-05-10 15:25:35.795714 PDT | [0] Epoch -208 finished +---------------------------------- --------------- +epoch -208 +replay_buffer/size 999033 +trainer/num train calls 793000 +trainer/Policy Loss -18.4955 +trainer/Log Pis Mean 23.7595 +trainer/Log Pis Std 13.4393 +trainer/Log Pis Max 80.569 +trainer/Log Pis Min -11.3284 +trainer/policy/mean Mean 0.00175602 +trainer/policy/mean Std 0.901486 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.72099 +trainer/policy/normal/std Std 0.670714 +trainer/policy/normal/std Max 6.49064 +trainer/policy/normal/std Min 0.290338 +trainer/policy/normal/log_std Mean 0.954657 +trainer/policy/normal/log_std Std 0.349206 +trainer/policy/normal/log_std Max 1.87036 +trainer/policy/normal/log_std Min -1.23671 +eval/num steps total 789320 +eval/num paths total 793 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109678 +eval/Actions Std 0.911699 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6744 +time/logging (s) 0.00368515 +time/sampling batch (s) 0.782936 +time/saving (s) 0.00346327 +time/training (s) 7.72221 +time/epoch (s) 11.1867 +time/total (s) 8087.98 +Epoch -208 +---------------------------------- --------------- +2022-05-10 15:25:45.341487 PDT | [0] Epoch -207 finished +---------------------------------- --------------- +epoch -207 +replay_buffer/size 999033 +trainer/num train calls 794000 +trainer/Policy Loss -19.7769 +trainer/Log Pis Mean 24.4126 +trainer/Log Pis Std 13.275 +trainer/Log Pis Max 60.3919 +trainer/Log Pis Min -7.88561 +trainer/policy/mean Mean -0.0425807 +trainer/policy/mean Std 0.907267 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.77377 +trainer/policy/normal/std Std 0.681507 +trainer/policy/normal/std Max 6.42835 +trainer/policy/normal/std Min 0.233588 +trainer/policy/normal/log_std Mean 0.97534 +trainer/policy/normal/log_std Std 0.342417 +trainer/policy/normal/log_std Max 1.86072 +trainer/policy/normal/log_std Min -1.45419 +eval/num steps total 790320 +eval/num paths total 794 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.167958 +eval/Actions Std 0.924171 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75409 +time/logging (s) 0.00381923 +time/sampling batch (s) 0.280213 +time/saving (s) 0.00344005 +time/training (s) 6.48108 +time/epoch (s) 9.52264 +time/total (s) 8097.51 +Epoch -207 +---------------------------------- --------------- +2022-05-10 15:25:55.421089 PDT | [0] Epoch -206 finished +---------------------------------- --------------- +epoch -206 +replay_buffer/size 999033 +trainer/num train calls 795000 +trainer/Policy Loss -20.0822 +trainer/Log Pis Mean 24.1961 +trainer/Log Pis Std 13.5848 +trainer/Log Pis Max 73.9113 +trainer/Log Pis Min -6.00801 +trainer/policy/mean Mean -0.0606566 +trainer/policy/mean Std 0.907266 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83039 +trainer/policy/normal/std Std 0.685922 +trainer/policy/normal/std Max 6.00305 +trainer/policy/normal/std Min 0.227863 +trainer/policy/normal/log_std Mean 0.996667 +trainer/policy/normal/log_std Std 0.337559 +trainer/policy/normal/log_std Max 1.79227 +trainer/policy/normal/log_std Min -1.47901 +eval/num steps total 791320 +eval/num paths total 795 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0630927 +eval/Actions Std 0.916854 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71975 +time/logging (s) 0.00378074 +time/sampling batch (s) 0.282512 +time/saving (s) 0.00344642 +time/training (s) 7.04663 +time/epoch (s) 10.0561 +time/total (s) 8107.57 +Epoch -206 +---------------------------------- --------------- +2022-05-10 15:26:05.476902 PDT | [0] Epoch -205 finished +---------------------------------- --------------- +epoch -205 +replay_buffer/size 999033 +trainer/num train calls 796000 +trainer/Policy Loss -19.4129 +trainer/Log Pis Mean 23.6545 +trainer/Log Pis Std 12.5371 +trainer/Log Pis Max 65.0689 +trainer/Log Pis Min -6.70735 +trainer/policy/mean Mean -0.0342255 +trainer/policy/mean Std 0.904112 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.79523 +trainer/policy/normal/std Std 0.651528 +trainer/policy/normal/std Max 5.75148 +trainer/policy/normal/std Min 0.275161 +trainer/policy/normal/log_std Mean 0.988077 +trainer/policy/normal/log_std Std 0.320474 +trainer/policy/normal/log_std Max 1.74946 +trainer/policy/normal/log_std Min -1.2904 +eval/num steps total 792320 +eval/num paths total 796 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.174813 +eval/Actions Std 0.881991 +eval/Actions Max 0.999999 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.82732 +time/logging (s) 0.00417879 +time/sampling batch (s) 0.287683 +time/saving (s) 0.00405783 +time/training (s) 6.909 +time/epoch (s) 10.0322 +time/total (s) 8117.6 +Epoch -205 +---------------------------------- --------------- +2022-05-10 15:26:15.849071 PDT | [0] Epoch -204 finished +---------------------------------- --------------- +epoch -204 +replay_buffer/size 999033 +trainer/num train calls 797000 +trainer/Policy Loss -20.5015 +trainer/Log Pis Mean 24.342 +trainer/Log Pis Std 12.8899 +trainer/Log Pis Max 75.4639 +trainer/Log Pis Min -8.30255 +trainer/policy/mean Mean -0.0478223 +trainer/policy/mean Std 0.911552 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78271 +trainer/policy/normal/std Std 0.661968 +trainer/policy/normal/std Max 6.42202 +trainer/policy/normal/std Min 0.205609 +trainer/policy/normal/log_std Mean 0.98075 +trainer/policy/normal/log_std Std 0.336023 +trainer/policy/normal/log_std Max 1.85973 +trainer/policy/normal/log_std Min -1.58178 +eval/num steps total 793320 +eval/num paths total 797 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0990723 +eval/Actions Std 0.899089 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5418 +time/logging (s) 0.00382428 +time/sampling batch (s) 0.528874 +time/saving (s) 0.00370529 +time/training (s) 7.27002 +time/epoch (s) 10.3482 +time/total (s) 8127.95 +Epoch -204 +---------------------------------- --------------- +2022-05-10 15:26:26.276301 PDT | [0] Epoch -203 finished +---------------------------------- --------------- +epoch -203 +replay_buffer/size 999033 +trainer/num train calls 798000 +trainer/Policy Loss -19.0172 +trainer/Log Pis Mean 24.174 +trainer/Log Pis Std 12.8323 +trainer/Log Pis Max 72.7901 +trainer/Log Pis Min -7.05288 +trainer/policy/mean Mean -0.0240235 +trainer/policy/mean Std 0.904889 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999975 +trainer/policy/normal/std Mean 2.78501 +trainer/policy/normal/std Std 0.660101 +trainer/policy/normal/std Max 5.50354 +trainer/policy/normal/std Min 0.284818 +trainer/policy/normal/log_std Mean 0.982301 +trainer/policy/normal/log_std Std 0.329975 +trainer/policy/normal/log_std Max 1.70539 +trainer/policy/normal/log_std Min -1.25591 +eval/num steps total 794320 +eval/num paths total 798 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0795329 +eval/Actions Std 0.915504 +eval/Actions Max 0.999991 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44994 +time/logging (s) 0.00377981 +time/sampling batch (s) 0.280849 +time/saving (s) 0.00356323 +time/training (s) 7.66559 +time/epoch (s) 10.4037 +time/total (s) 8138.36 +Epoch -203 +---------------------------------- --------------- +2022-05-10 15:26:35.997380 PDT | [0] Epoch -202 finished +---------------------------------- --------------- +epoch -202 +replay_buffer/size 999033 +trainer/num train calls 799000 +trainer/Policy Loss -20.1781 +trainer/Log Pis Mean 25.2465 +trainer/Log Pis Std 13.6991 +trainer/Log Pis Max 66.7449 +trainer/Log Pis Min -11.6324 +trainer/policy/mean Mean -0.0307038 +trainer/policy/mean Std 0.910643 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80633 +trainer/policy/normal/std Std 0.683352 +trainer/policy/normal/std Max 5.83845 +trainer/policy/normal/std Min 0.263884 +trainer/policy/normal/log_std Mean 0.987962 +trainer/policy/normal/log_std Std 0.33736 +trainer/policy/normal/log_std Max 1.76447 +trainer/policy/normal/log_std Min -1.33225 +eval/num steps total 795320 +eval/num paths total 799 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.469878 +eval/Actions Std 0.670366 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64256 +time/logging (s) 0.00367276 +time/sampling batch (s) 0.278632 +time/saving (s) 0.00345621 +time/training (s) 6.76936 +time/epoch (s) 9.69767 +time/total (s) 8148.06 +Epoch -202 +---------------------------------- --------------- +2022-05-10 15:26:45.430589 PDT | [0] Epoch -201 finished +---------------------------------- --------------- +epoch -201 +replay_buffer/size 999033 +trainer/num train calls 800000 +trainer/Policy Loss -19.7163 +trainer/Log Pis Mean 24.282 +trainer/Log Pis Std 13.3219 +trainer/Log Pis Max 66.8302 +trainer/Log Pis Min -5.39985 +trainer/policy/mean Mean -0.0405186 +trainer/policy/mean Std 0.904442 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.78841 +trainer/policy/normal/std Std 0.656369 +trainer/policy/normal/std Max 5.44692 +trainer/policy/normal/std Min 0.289205 +trainer/policy/normal/log_std Mean 0.984407 +trainer/policy/normal/log_std Std 0.326212 +trainer/policy/normal/log_std Max 1.69505 +trainer/policy/normal/log_std Min -1.24062 +eval/num steps total 796320 +eval/num paths total 800 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.161767 +eval/Actions Std 0.861338 +eval/Actions Max 0.999997 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63085 +time/logging (s) 0.0037258 +time/sampling batch (s) 0.278847 +time/saving (s) 0.00633999 +time/training (s) 6.49017 +time/epoch (s) 9.40993 +time/total (s) 8157.47 +Epoch -201 +---------------------------------- --------------- +2022-05-10 15:26:55.429530 PDT | [0] Epoch -200 finished +---------------------------------- --------------- +epoch -200 +replay_buffer/size 999033 +trainer/num train calls 801000 +trainer/Policy Loss -20.4453 +trainer/Log Pis Mean 24.7973 +trainer/Log Pis Std 13.7193 +trainer/Log Pis Max 71.0115 +trainer/Log Pis Min -9.00875 +trainer/policy/mean Mean -0.0450359 +trainer/policy/mean Std 0.907706 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.78257 +trainer/policy/normal/std Std 0.680952 +trainer/policy/normal/std Max 5.31539 +trainer/policy/normal/std Min 0.22886 +trainer/policy/normal/log_std Mean 0.978775 +trainer/policy/normal/log_std Std 0.339899 +trainer/policy/normal/log_std Max 1.67061 +trainer/policy/normal/log_std Min -1.47464 +eval/num steps total 797320 +eval/num paths total 801 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118987 +eval/Actions Std 0.912719 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.602 +time/logging (s) 0.003757 +time/sampling batch (s) 0.279947 +time/saving (s) 0.00343622 +time/training (s) 7.08643 +time/epoch (s) 9.97557 +time/total (s) 8167.45 +Epoch -200 +---------------------------------- --------------- +2022-05-10 15:27:06.315761 PDT | [0] Epoch -199 finished +---------------------------------- --------------- +epoch -199 +replay_buffer/size 999033 +trainer/num train calls 802000 +trainer/Policy Loss -20.8213 +trainer/Log Pis Mean 24.5066 +trainer/Log Pis Std 13.8762 +trainer/Log Pis Max 85.4217 +trainer/Log Pis Min -9.17591 +trainer/policy/mean Mean -0.0136661 +trainer/policy/mean Std 0.911196 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81398 +trainer/policy/normal/std Std 0.659776 +trainer/policy/normal/std Max 5.72806 +trainer/policy/normal/std Min 0.236661 +trainer/policy/normal/log_std Mean 0.994031 +trainer/policy/normal/log_std Std 0.324381 +trainer/policy/normal/log_std Max 1.74538 +trainer/policy/normal/log_std Min -1.44113 +eval/num steps total 798320 +eval/num paths total 802 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.351782 +eval/Actions Std 0.815311 +eval/Actions Max 0.999983 +eval/Actions Min -0.999969 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6939 +time/logging (s) 0.003684 +time/sampling batch (s) 0.281538 +time/saving (s) 0.00343883 +time/training (s) 7.88013 +time/epoch (s) 10.8627 +time/total (s) 8178.32 +Epoch -199 +---------------------------------- --------------- +2022-05-10 15:27:16.255508 PDT | [0] Epoch -198 finished +---------------------------------- --------------- +epoch -198 +replay_buffer/size 999033 +trainer/num train calls 803000 +trainer/Policy Loss -20.2338 +trainer/Log Pis Mean 24.1525 +trainer/Log Pis Std 12.5946 +trainer/Log Pis Max 68.7958 +trainer/Log Pis Min -7.2853 +trainer/policy/mean Mean -0.0364025 +trainer/policy/mean Std 0.90804 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.87609 +trainer/policy/normal/std Std 0.67726 +trainer/policy/normal/std Max 6.01679 +trainer/policy/normal/std Min 0.261499 +trainer/policy/normal/log_std Mean 1.01477 +trainer/policy/normal/log_std Std 0.330809 +trainer/policy/normal/log_std Max 1.79455 +trainer/policy/normal/log_std Min -1.34132 +eval/num steps total 799320 +eval/num paths total 803 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0804416 +eval/Actions Std 0.909996 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81245 +time/logging (s) 0.00415261 +time/sampling batch (s) 0.281877 +time/saving (s) 0.00410998 +time/training (s) 6.81418 +time/epoch (s) 9.91677 +time/total (s) 8188.24 +Epoch -198 +---------------------------------- --------------- +2022-05-10 15:27:26.558207 PDT | [0] Epoch -197 finished +---------------------------------- --------------- +epoch -197 +replay_buffer/size 999033 +trainer/num train calls 804000 +trainer/Policy Loss -19.544 +trainer/Log Pis Mean 24.9251 +trainer/Log Pis Std 13.3444 +trainer/Log Pis Max 64.9022 +trainer/Log Pis Min -7.76604 +trainer/policy/mean Mean -0.0343788 +trainer/policy/mean Std 0.908174 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.81028 +trainer/policy/normal/std Std 0.650315 +trainer/policy/normal/std Max 5.49555 +trainer/policy/normal/std Min 0.278463 +trainer/policy/normal/log_std Mean 0.994965 +trainer/policy/normal/log_std Std 0.311753 +trainer/policy/normal/log_std Max 1.70394 +trainer/policy/normal/log_std Min -1.27847 +eval/num steps total 800320 +eval/num paths total 804 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.123041 +eval/Actions Std 0.914602 +eval/Actions Max 1 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52824 +time/logging (s) 0.0037278 +time/sampling batch (s) 0.280932 +time/saving (s) 0.00359149 +time/training (s) 7.46175 +time/epoch (s) 10.2782 +time/total (s) 8198.52 +Epoch -197 +---------------------------------- --------------- +2022-05-10 15:27:37.456826 PDT | [0] Epoch -196 finished +---------------------------------- --------------- +epoch -196 +replay_buffer/size 999033 +trainer/num train calls 805000 +trainer/Policy Loss -19.0216 +trainer/Log Pis Mean 24.6852 +trainer/Log Pis Std 12.7568 +trainer/Log Pis Max 75.5691 +trainer/Log Pis Min -3.38097 +trainer/policy/mean Mean -0.0474481 +trainer/policy/mean Std 0.905459 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.82761 +trainer/policy/normal/std Std 0.664595 +trainer/policy/normal/std Max 5.71893 +trainer/policy/normal/std Min 0.286358 +trainer/policy/normal/log_std Mean 0.998912 +trainer/policy/normal/log_std Std 0.323848 +trainer/policy/normal/log_std Max 1.74378 +trainer/policy/normal/log_std Min -1.25051 +eval/num steps total 801320 +eval/num paths total 805 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0173015 +eval/Actions Std 0.874266 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44746 +time/logging (s) 0.00374802 +time/sampling batch (s) 0.28476 +time/saving (s) 0.00361196 +time/training (s) 8.13519 +time/epoch (s) 10.8748 +time/total (s) 8209.4 +Epoch -196 +---------------------------------- --------------- +2022-05-10 15:27:47.807791 PDT | [0] Epoch -195 finished +---------------------------------- --------------- +epoch -195 +replay_buffer/size 999033 +trainer/num train calls 806000 +trainer/Policy Loss -19.9912 +trainer/Log Pis Mean 23.9662 +trainer/Log Pis Std 13.0543 +trainer/Log Pis Max 68.3268 +trainer/Log Pis Min -7.98777 +trainer/policy/mean Mean -0.0381173 +trainer/policy/mean Std 0.90489 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76 +trainer/policy/normal/std Std 0.679802 +trainer/policy/normal/std Max 6.29463 +trainer/policy/normal/std Min 0.241986 +trainer/policy/normal/log_std Mean 0.97137 +trainer/policy/normal/log_std Std 0.334763 +trainer/policy/normal/log_std Max 1.8397 +trainer/policy/normal/log_std Min -1.41888 +eval/num steps total 802320 +eval/num paths total 806 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178318 +eval/Actions Std 0.889923 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.86595 +time/logging (s) 0.00370067 +time/sampling batch (s) 0.280573 +time/saving (s) 0.0034576 +time/training (s) 7.17379 +time/epoch (s) 10.3275 +time/total (s) 8219.73 +Epoch -195 +---------------------------------- --------------- +2022-05-10 15:27:57.641836 PDT | [0] Epoch -194 finished +---------------------------------- --------------- +epoch -194 +replay_buffer/size 999033 +trainer/num train calls 807000 +trainer/Policy Loss -20.375 +trainer/Log Pis Mean 24.5805 +trainer/Log Pis Std 13.2587 +trainer/Log Pis Max 66.7024 +trainer/Log Pis Min -7.64184 +trainer/policy/mean Mean -0.0275449 +trainer/policy/mean Std 0.907119 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75726 +trainer/policy/normal/std Std 0.700308 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.228704 +trainer/policy/normal/log_std Mean 0.964128 +trainer/policy/normal/log_std Std 0.365272 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.47533 +eval/num steps total 803320 +eval/num paths total 807 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0880591 +eval/Actions Std 0.869341 +eval/Actions Max 0.999999 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61744 +time/logging (s) 0.00368772 +time/sampling batch (s) 0.277363 +time/saving (s) 0.00343254 +time/training (s) 6.90892 +time/epoch (s) 9.81084 +time/total (s) 8229.54 +Epoch -194 +---------------------------------- --------------- +2022-05-10 15:28:06.980462 PDT | [0] Epoch -193 finished +---------------------------------- --------------- +epoch -193 +replay_buffer/size 999033 +trainer/num train calls 808000 +trainer/Policy Loss -20.491 +trainer/Log Pis Mean 23.4634 +trainer/Log Pis Std 13.6537 +trainer/Log Pis Max 75.7658 +trainer/Log Pis Min -10.2074 +trainer/policy/mean Mean -0.0207511 +trainer/policy/mean Std 0.905875 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8281 +trainer/policy/normal/std Std 0.688575 +trainer/policy/normal/std Max 6.67769 +trainer/policy/normal/std Min 0.268712 +trainer/policy/normal/log_std Mean 0.995349 +trainer/policy/normal/log_std Std 0.339753 +trainer/policy/normal/log_std Max 1.89877 +trainer/policy/normal/log_std Min -1.31412 +eval/num steps total 804320 +eval/num paths total 808 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.192352 +eval/Actions Std 0.851676 +eval/Actions Max 0.999994 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45935 +time/logging (s) 0.00371666 +time/sampling batch (s) 0.284209 +time/saving (s) 0.00353412 +time/training (s) 6.56415 +time/epoch (s) 9.31496 +time/total (s) 8238.86 +Epoch -193 +---------------------------------- --------------- +2022-05-10 15:28:16.711920 PDT | [0] Epoch -192 finished +---------------------------------- --------------- +epoch -192 +replay_buffer/size 999033 +trainer/num train calls 809000 +trainer/Policy Loss -20.8573 +trainer/Log Pis Mean 25.148 +trainer/Log Pis Std 13.6285 +trainer/Log Pis Max 67.994 +trainer/Log Pis Min -10.2209 +trainer/policy/mean Mean -0.030162 +trainer/policy/mean Std 0.908871 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81833 +trainer/policy/normal/std Std 0.639506 +trainer/policy/normal/std Max 5.4666 +trainer/policy/normal/std Min 0.284371 +trainer/policy/normal/log_std Mean 0.999205 +trainer/policy/normal/log_std Std 0.305864 +trainer/policy/normal/log_std Max 1.69866 +trainer/policy/normal/log_std Min -1.25747 +eval/num steps total 805320 +eval/num paths total 809 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.282942 +eval/Actions Std 0.884968 +eval/Actions Max 0.999998 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60098 +time/logging (s) 0.00436937 +time/sampling batch (s) 0.283746 +time/saving (s) 0.00443124 +time/training (s) 6.81474 +time/epoch (s) 9.70827 +time/total (s) 8248.57 +Epoch -192 +---------------------------------- --------------- +2022-05-10 15:28:26.192650 PDT | [0] Epoch -191 finished +---------------------------------- --------------- +epoch -191 +replay_buffer/size 999033 +trainer/num train calls 810000 +trainer/Policy Loss -20.7353 +trainer/Log Pis Mean 24.7929 +trainer/Log Pis Std 13.4735 +trainer/Log Pis Max 65.6241 +trainer/Log Pis Min -9.40968 +trainer/policy/mean Mean -0.0205853 +trainer/policy/mean Std 0.908779 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.78819 +trainer/policy/normal/std Std 0.64059 +trainer/policy/normal/std Max 7.3166 +trainer/policy/normal/std Min 0.251599 +trainer/policy/normal/log_std Mean 0.987698 +trainer/policy/normal/log_std Std 0.30892 +trainer/policy/normal/log_std Max 1.99015 +trainer/policy/normal/log_std Min -1.37992 +eval/num steps total 806320 +eval/num paths total 810 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.239882 +eval/Actions Std 0.932314 +eval/Actions Max 0.999987 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6222 +time/logging (s) 0.0038049 +time/sampling batch (s) 0.279999 +time/saving (s) 0.00365949 +time/training (s) 6.54531 +time/epoch (s) 9.45497 +time/total (s) 8258.03 +Epoch -191 +---------------------------------- --------------- +2022-05-10 15:28:36.154883 PDT | [0] Epoch -190 finished +---------------------------------- --------------- +epoch -190 +replay_buffer/size 999033 +trainer/num train calls 811000 +trainer/Policy Loss -20.4443 +trainer/Log Pis Mean 23.7277 +trainer/Log Pis Std 12.8202 +trainer/Log Pis Max 66.4385 +trainer/Log Pis Min -9.38918 +trainer/policy/mean Mean -0.0322044 +trainer/policy/mean Std 0.908422 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77992 +trainer/policy/normal/std Std 0.673044 +trainer/policy/normal/std Max 6.16361 +trainer/policy/normal/std Min 0.267659 +trainer/policy/normal/log_std Mean 0.979447 +trainer/policy/normal/log_std Std 0.332577 +trainer/policy/normal/log_std Max 1.81866 +trainer/policy/normal/log_std Min -1.31804 +eval/num steps total 807320 +eval/num paths total 811 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.108117 +eval/Actions Std 0.885099 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54757 +time/logging (s) 0.00385603 +time/sampling batch (s) 0.27842 +time/saving (s) 0.00342041 +time/training (s) 7.10556 +time/epoch (s) 9.93882 +time/total (s) 8267.97 +Epoch -190 +---------------------------------- --------------- +2022-05-10 15:28:46.726138 PDT | [0] Epoch -189 finished +---------------------------------- --------------- +epoch -189 +replay_buffer/size 999033 +trainer/num train calls 812000 +trainer/Policy Loss -19.4009 +trainer/Log Pis Mean 23.8246 +trainer/Log Pis Std 13.2521 +trainer/Log Pis Max 75.39 +trainer/Log Pis Min -8.81637 +trainer/policy/mean Mean -0.0211233 +trainer/policy/mean Std 0.907695 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.71215 +trainer/policy/normal/std Std 0.64369 +trainer/policy/normal/std Max 5.49951 +trainer/policy/normal/std Min 0.313935 +trainer/policy/normal/log_std Mean 0.956907 +trainer/policy/normal/log_std Std 0.323136 +trainer/policy/normal/log_std Max 1.70466 +trainer/policy/normal/log_std Min -1.15857 +eval/num steps total 808320 +eval/num paths total 812 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.115313 +eval/Actions Std 0.914073 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58042 +time/logging (s) 0.00380429 +time/sampling batch (s) 0.281519 +time/saving (s) 0.00346149 +time/training (s) 7.67851 +time/epoch (s) 10.5477 +time/total (s) 8278.52 +Epoch -189 +---------------------------------- --------------- +2022-05-10 15:28:57.070468 PDT | [0] Epoch -188 finished +---------------------------------- --------------- +epoch -188 +replay_buffer/size 999033 +trainer/num train calls 813000 +trainer/Policy Loss -20.4113 +trainer/Log Pis Mean 24.7463 +trainer/Log Pis Std 12.8733 +trainer/Log Pis Max 71.5931 +trainer/Log Pis Min -9.21078 +trainer/policy/mean Mean -0.04138 +trainer/policy/mean Std 0.905673 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83129 +trainer/policy/normal/std Std 0.689413 +trainer/policy/normal/std Max 5.40657 +trainer/policy/normal/std Min 0.269095 +trainer/policy/normal/log_std Mean 0.996475 +trainer/policy/normal/log_std Std 0.33997 +trainer/policy/normal/log_std Max 1.68762 +trainer/policy/normal/log_std Min -1.31269 +eval/num steps total 809320 +eval/num paths total 813 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00257771 +eval/Actions Std 0.917792 +eval/Actions Max 0.999994 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60453 +time/logging (s) 0.0036906 +time/sampling batch (s) 0.52885 +time/saving (s) 0.00344387 +time/training (s) 7.18057 +time/epoch (s) 10.3211 +time/total (s) 8288.84 +Epoch -188 +---------------------------------- --------------- +2022-05-10 15:29:07.633007 PDT | [0] Epoch -187 finished +---------------------------------- --------------- +epoch -187 +replay_buffer/size 999033 +trainer/num train calls 814000 +trainer/Policy Loss -19.478 +trainer/Log Pis Mean 23.9371 +trainer/Log Pis Std 12.2159 +trainer/Log Pis Max 67.4657 +trainer/Log Pis Min -10.8021 +trainer/policy/mean Mean -0.0401476 +trainer/policy/mean Std 0.899008 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76358 +trainer/policy/normal/std Std 0.694962 +trainer/policy/normal/std Max 6.96495 +trainer/policy/normal/std Min 0.304305 +trainer/policy/normal/log_std Mean 0.969511 +trainer/policy/normal/log_std Std 0.348829 +trainer/policy/normal/log_std Max 1.94089 +trainer/policy/normal/log_std Min -1.18972 +eval/num steps total 810320 +eval/num paths total 814 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0188451 +eval/Actions Std 0.909859 +eval/Actions Max 0.999997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59181 +time/logging (s) 0.00376472 +time/sampling batch (s) 0.283733 +time/saving (s) 0.00342668 +time/training (s) 7.65645 +time/epoch (s) 10.5392 +time/total (s) 8299.39 +Epoch -187 +---------------------------------- --------------- +2022-05-10 15:29:17.939702 PDT | [0] Epoch -186 finished +---------------------------------- --------------- +epoch -186 +replay_buffer/size 999033 +trainer/num train calls 815000 +trainer/Policy Loss -20.6704 +trainer/Log Pis Mean 25.0116 +trainer/Log Pis Std 13.2138 +trainer/Log Pis Max 67.3165 +trainer/Log Pis Min -3.34853 +trainer/policy/mean Mean -0.0406676 +trainer/policy/mean Std 0.905968 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77186 +trainer/policy/normal/std Std 0.650876 +trainer/policy/normal/std Max 7.00631 +trainer/policy/normal/std Min 0.210344 +trainer/policy/normal/log_std Mean 0.978806 +trainer/policy/normal/log_std Std 0.325921 +trainer/policy/normal/log_std Max 1.94681 +trainer/policy/normal/log_std Min -1.55901 +eval/num steps total 811320 +eval/num paths total 815 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0286368 +eval/Actions Std 0.908217 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59442 +time/logging (s) 0.00370477 +time/sampling batch (s) 0.27862 +time/saving (s) 0.00343825 +time/training (s) 7.40316 +time/epoch (s) 10.2833 +time/total (s) 8309.67 +Epoch -186 +---------------------------------- --------------- +2022-05-10 15:29:27.760177 PDT | [0] Epoch -185 finished +---------------------------------- --------------- +epoch -185 +replay_buffer/size 999033 +trainer/num train calls 816000 +trainer/Policy Loss -19.3814 +trainer/Log Pis Mean 23.848 +trainer/Log Pis Std 13.2765 +trainer/Log Pis Max 71.2938 +trainer/Log Pis Min -5.48076 +trainer/policy/mean Mean -0.0501163 +trainer/policy/mean Std 0.905789 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.76782 +trainer/policy/normal/std Std 0.677395 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.277046 +trainer/policy/normal/log_std Mean 0.973522 +trainer/policy/normal/log_std Std 0.340861 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.28357 +eval/num steps total 812320 +eval/num paths total 816 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.160299 +eval/Actions Std 0.927589 +eval/Actions Max 0.999993 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53059 +time/logging (s) 0.00372631 +time/sampling batch (s) 0.282212 +time/saving (s) 0.00358723 +time/training (s) 6.9769 +time/epoch (s) 9.79701 +time/total (s) 8319.47 +Epoch -185 +---------------------------------- --------------- +2022-05-10 15:29:36.577166 PDT | [0] Epoch -184 finished +---------------------------------- --------------- +epoch -184 +replay_buffer/size 999033 +trainer/num train calls 817000 +trainer/Policy Loss -18.6528 +trainer/Log Pis Mean 25.5035 +trainer/Log Pis Std 13.2241 +trainer/Log Pis Max 73.6326 +trainer/Log Pis Min -14.0083 +trainer/policy/mean Mean -0.0514684 +trainer/policy/mean Std 0.906162 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.8505 +trainer/policy/normal/std Std 0.682434 +trainer/policy/normal/std Max 5.95531 +trainer/policy/normal/std Min 0.28957 +trainer/policy/normal/log_std Mean 1.00586 +trainer/policy/normal/log_std Std 0.327181 +trainer/policy/normal/log_std Max 1.78428 +trainer/policy/normal/log_std Min -1.23936 +eval/num steps total 813320 +eval/num paths total 817 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.017103 +eval/Actions Std 0.904106 +eval/Actions Max 0.999991 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48675 +time/logging (s) 0.00365025 +time/sampling batch (s) 0.284453 +time/saving (s) 0.00346025 +time/training (s) 6.01475 +time/epoch (s) 8.79307 +time/total (s) 8328.27 +Epoch -184 +---------------------------------- --------------- +2022-05-10 15:29:45.731864 PDT | [0] Epoch -183 finished +---------------------------------- --------------- +epoch -183 +replay_buffer/size 999033 +trainer/num train calls 818000 +trainer/Policy Loss -19.9134 +trainer/Log Pis Mean 24.7808 +trainer/Log Pis Std 13.407 +trainer/Log Pis Max 60.841 +trainer/Log Pis Min -9.52085 +trainer/policy/mean Mean -0.0159685 +trainer/policy/mean Std 0.907037 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.84184 +trainer/policy/normal/std Std 0.68017 +trainer/policy/normal/std Max 6.59837 +trainer/policy/normal/std Min 0.275758 +trainer/policy/normal/log_std Mean 1.0041 +trainer/policy/normal/log_std Std 0.319593 +trainer/policy/normal/log_std Max 1.88682 +trainer/policy/normal/log_std Min -1.28823 +eval/num steps total 814320 +eval/num paths total 818 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.119983 +eval/Actions Std 0.904787 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68536 +time/logging (s) 0.00420698 +time/sampling batch (s) 0.280866 +time/saving (s) 0.00408648 +time/training (s) 6.15703 +time/epoch (s) 9.13154 +time/total (s) 8337.4 +Epoch -183 +---------------------------------- --------------- +2022-05-10 15:29:55.924566 PDT | [0] Epoch -182 finished +---------------------------------- --------------- +epoch -182 +replay_buffer/size 999033 +trainer/num train calls 819000 +trainer/Policy Loss -20.851 +trainer/Log Pis Mean 24.6863 +trainer/Log Pis Std 13.8234 +trainer/Log Pis Max 75.722 +trainer/Log Pis Min -11.5566 +trainer/policy/mean Mean -0.0370644 +trainer/policy/mean Std 0.913248 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81412 +trainer/policy/normal/std Std 0.645046 +trainer/policy/normal/std Max 5.45041 +trainer/policy/normal/std Min 0.27249 +trainer/policy/normal/log_std Mean 0.996164 +trainer/policy/normal/log_std Std 0.314848 +trainer/policy/normal/log_std Max 1.69569 +trainer/policy/normal/log_std Min -1.30015 +eval/num steps total 815320 +eval/num paths total 819 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222682 +eval/Actions Std 0.924367 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7335 +time/logging (s) 0.00385037 +time/sampling batch (s) 0.530407 +time/saving (s) 0.00372188 +time/training (s) 6.89695 +time/epoch (s) 10.1684 +time/total (s) 8347.57 +Epoch -182 +---------------------------------- --------------- +2022-05-10 15:30:05.823468 PDT | [0] Epoch -181 finished +---------------------------------- --------------- +epoch -181 +replay_buffer/size 999033 +trainer/num train calls 820000 +trainer/Policy Loss -19.7479 +trainer/Log Pis Mean 24.9077 +trainer/Log Pis Std 13.1869 +trainer/Log Pis Max 66.154 +trainer/Log Pis Min -7.11293 +trainer/policy/mean Mean -0.00952468 +trainer/policy/mean Std 0.907499 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78234 +trainer/policy/normal/std Std 0.652264 +trainer/policy/normal/std Max 5.39183 +trainer/policy/normal/std Min 0.324196 +trainer/policy/normal/log_std Mean 0.983197 +trainer/policy/normal/log_std Std 0.321396 +trainer/policy/normal/log_std Max 1.68489 +trainer/policy/normal/log_std Min -1.12641 +eval/num steps total 816320 +eval/num paths total 820 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.243589 +eval/Actions Std 0.885227 +eval/Actions Max 0.999996 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75466 +time/logging (s) 0.00367293 +time/sampling batch (s) 0.282529 +time/saving (s) 0.00345251 +time/training (s) 6.8306 +time/epoch (s) 9.87492 +time/total (s) 8357.45 +Epoch -181 +---------------------------------- --------------- +2022-05-10 15:30:15.474352 PDT | [0] Epoch -180 finished +---------------------------------- --------------- +epoch -180 +replay_buffer/size 999033 +trainer/num train calls 821000 +trainer/Policy Loss -20.6386 +trainer/Log Pis Mean 24.0249 +trainer/Log Pis Std 12.2003 +trainer/Log Pis Max 62.1173 +trainer/Log Pis Min -6.9393 +trainer/policy/mean Mean -0.0603322 +trainer/policy/mean Std 0.909603 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.75297 +trainer/policy/normal/std Std 0.637925 +trainer/policy/normal/std Max 5.85308 +trainer/policy/normal/std Min 0.309617 +trainer/policy/normal/log_std Mean 0.974004 +trainer/policy/normal/log_std Std 0.314031 +trainer/policy/normal/log_std Max 1.76697 +trainer/policy/normal/log_std Min -1.17242 +eval/num steps total 817320 +eval/num paths total 821 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.277814 +eval/Actions Std 0.825139 +eval/Actions Max 0.999977 +eval/Actions Min -0.999955 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45418 +time/logging (s) 0.00367122 +time/sampling batch (s) 0.530281 +time/saving (s) 0.00346235 +time/training (s) 6.63596 +time/epoch (s) 9.62756 +time/total (s) 8367.08 +Epoch -180 +---------------------------------- --------------- +2022-05-10 15:30:26.030689 PDT | [0] Epoch -179 finished +---------------------------------- --------------- +epoch -179 +replay_buffer/size 999033 +trainer/num train calls 822000 +trainer/Policy Loss -20.0196 +trainer/Log Pis Mean 24.3574 +trainer/Log Pis Std 13.0777 +trainer/Log Pis Max 64.7039 +trainer/Log Pis Min -6.14494 +trainer/policy/mean Mean -0.0137853 +trainer/policy/mean Std 0.907007 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8011 +trainer/policy/normal/std Std 0.671117 +trainer/policy/normal/std Max 6.56844 +trainer/policy/normal/std Min 0.297759 +trainer/policy/normal/log_std Mean 0.988672 +trainer/policy/normal/log_std Std 0.325258 +trainer/policy/normal/log_std Max 1.88228 +trainer/policy/normal/log_std Min -1.21147 +eval/num steps total 818320 +eval/num paths total 822 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109239 +eval/Actions Std 0.900959 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52811 +time/logging (s) 0.00370081 +time/sampling batch (s) 0.282703 +time/saving (s) 0.0034928 +time/training (s) 7.71497 +time/epoch (s) 10.533 +time/total (s) 8377.62 +Epoch -179 +---------------------------------- --------------- +2022-05-10 15:30:36.323420 PDT | [0] Epoch -178 finished +---------------------------------- --------------- +epoch -178 +replay_buffer/size 999033 +trainer/num train calls 823000 +trainer/Policy Loss -18.8156 +trainer/Log Pis Mean 24.7024 +trainer/Log Pis Std 13.6067 +trainer/Log Pis Max 68.1266 +trainer/Log Pis Min -7.27157 +trainer/policy/mean Mean -0.0194017 +trainer/policy/mean Std 0.902477 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84419 +trainer/policy/normal/std Std 0.700346 +trainer/policy/normal/std Max 6.48192 +trainer/policy/normal/std Min 0.246741 +trainer/policy/normal/log_std Mean 0.999183 +trainer/policy/normal/log_std Std 0.348532 +trainer/policy/normal/log_std Max 1.86902 +trainer/policy/normal/log_std Min -1.39942 +eval/num steps total 819320 +eval/num paths total 823 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.180095 +eval/Actions Std 0.870917 +eval/Actions Max 0.999998 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52213 +time/logging (s) 0.00418922 +time/sampling batch (s) 0.284072 +time/saving (s) 0.00424774 +time/training (s) 7.45467 +time/epoch (s) 10.2693 +time/total (s) 8387.89 +Epoch -178 +---------------------------------- --------------- +2022-05-10 15:30:47.052431 PDT | [0] Epoch -177 finished +---------------------------------- --------------- +epoch -177 +replay_buffer/size 999033 +trainer/num train calls 824000 +trainer/Policy Loss -19.5796 +trainer/Log Pis Mean 24.1949 +trainer/Log Pis Std 12.9994 +trainer/Log Pis Max 72.6178 +trainer/Log Pis Min -9.4271 +trainer/policy/mean Mean -0.03198 +trainer/policy/mean Std 0.909132 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.87196 +trainer/policy/normal/std Std 0.695101 +trainer/policy/normal/std Max 6.57875 +trainer/policy/normal/std Min 0.294679 +trainer/policy/normal/log_std Mean 1.01218 +trainer/policy/normal/log_std Std 0.332097 +trainer/policy/normal/log_std Max 1.88385 +trainer/policy/normal/log_std Min -1.22187 +eval/num steps total 820320 +eval/num paths total 824 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.189639 +eval/Actions Std 0.940417 +eval/Actions Max 0.999976 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51412 +time/logging (s) 0.00376948 +time/sampling batch (s) 0.283175 +time/saving (s) 0.00363598 +time/training (s) 7.89976 +time/epoch (s) 10.7045 +time/total (s) 8398.6 +Epoch -177 +---------------------------------- --------------- +2022-05-10 15:30:57.846622 PDT | [0] Epoch -176 finished +---------------------------------- --------------- +epoch -176 +replay_buffer/size 999033 +trainer/num train calls 825000 +trainer/Policy Loss -19.2553 +trainer/Log Pis Mean 24.1752 +trainer/Log Pis Std 13.2403 +trainer/Log Pis Max 74.6471 +trainer/Log Pis Min -9.78307 +trainer/policy/mean Mean -0.0364625 +trainer/policy/mean Std 0.903975 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.80739 +trainer/policy/normal/std Std 0.673823 +trainer/policy/normal/std Max 5.63245 +trainer/policy/normal/std Min 0.288523 +trainer/policy/normal/log_std Mean 0.989861 +trainer/policy/normal/log_std Std 0.331299 +trainer/policy/normal/log_std Max 1.72854 +trainer/policy/normal/log_std Min -1.24298 +eval/num steps total 821320 +eval/num paths total 825 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101482 +eval/Actions Std 0.896576 +eval/Actions Max 1 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71109 +time/logging (s) 0.00365215 +time/sampling batch (s) 0.281482 +time/saving (s) 0.00344586 +time/training (s) 7.77066 +time/epoch (s) 10.7703 +time/total (s) 8409.37 +Epoch -176 +---------------------------------- --------------- +2022-05-10 15:31:08.733193 PDT | [0] Epoch -175 finished +---------------------------------- --------------- +epoch -175 +replay_buffer/size 999033 +trainer/num train calls 826000 +trainer/Policy Loss -20.4911 +trainer/Log Pis Mean 25.2108 +trainer/Log Pis Std 13.8724 +trainer/Log Pis Max 76.5303 +trainer/Log Pis Min -10.5018 +trainer/policy/mean Mean -0.0478015 +trainer/policy/mean Std 0.906582 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.86058 +trainer/policy/normal/std Std 0.704087 +trainer/policy/normal/std Max 5.92537 +trainer/policy/normal/std Min 0.304338 +trainer/policy/normal/log_std Mean 1.00579 +trainer/policy/normal/log_std Std 0.343014 +trainer/policy/normal/log_std Max 1.77924 +trainer/policy/normal/log_std Min -1.18962 +eval/num steps total 822320 +eval/num paths total 826 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.236329 +eval/Actions Std 0.828601 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59707 +time/logging (s) 0.00372613 +time/sampling batch (s) 0.281626 +time/saving (s) 0.00345829 +time/training (s) 7.97722 +time/epoch (s) 10.8631 +time/total (s) 8420.24 +Epoch -175 +---------------------------------- --------------- +2022-05-10 15:31:18.779777 PDT | [0] Epoch -174 finished +---------------------------------- --------------- +epoch -174 +replay_buffer/size 999033 +trainer/num train calls 827000 +trainer/Policy Loss -20.0464 +trainer/Log Pis Mean 24.811 +trainer/Log Pis Std 13.8886 +trainer/Log Pis Max 70.7809 +trainer/Log Pis Min -8.13872 +trainer/policy/mean Mean -0.0178749 +trainer/policy/mean Std 0.90869 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.87094 +trainer/policy/normal/std Std 0.687219 +trainer/policy/normal/std Max 6.10661 +trainer/policy/normal/std Min 0.253793 +trainer/policy/normal/log_std Mean 1.01304 +trainer/policy/normal/log_std Std 0.326171 +trainer/policy/normal/log_std Max 1.80937 +trainer/policy/normal/log_std Min -1.37124 +eval/num steps total 823320 +eval/num paths total 827 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106005 +eval/Actions Std 0.916332 +eval/Actions Max 0.999996 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67901 +time/logging (s) 0.00371491 +time/sampling batch (s) 0.282422 +time/saving (s) 0.00347558 +time/training (s) 7.05427 +time/epoch (s) 10.0229 +time/total (s) 8430.26 +Epoch -174 +---------------------------------- --------------- +2022-05-10 15:31:28.096031 PDT | [0] Epoch -173 finished +---------------------------------- --------------- +epoch -173 +replay_buffer/size 999033 +trainer/num train calls 828000 +trainer/Policy Loss -18.9027 +trainer/Log Pis Mean 25.0493 +trainer/Log Pis Std 12.4069 +trainer/Log Pis Max 60.1453 +trainer/Log Pis Min -7.71255 +trainer/policy/mean Mean -0.0236461 +trainer/policy/mean Std 0.904754 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.82047 +trainer/policy/normal/std Std 0.67876 +trainer/policy/normal/std Max 5.40429 +trainer/policy/normal/std Min 0.20655 +trainer/policy/normal/log_std Mean 0.993606 +trainer/policy/normal/log_std Std 0.337308 +trainer/policy/normal/log_std Max 1.68719 +trainer/policy/normal/log_std Min -1.57721 +eval/num steps total 824320 +eval/num paths total 828 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.122087 +eval/Actions Std 0.913951 +eval/Actions Max 0.999999 +eval/Actions Min -0.999968 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64325 +time/logging (s) 0.00421711 +time/sampling batch (s) 0.283626 +time/saving (s) 0.00422084 +time/training (s) 6.35763 +time/epoch (s) 9.29294 +time/total (s) 8439.56 +Epoch -173 +---------------------------------- --------------- +2022-05-10 15:31:39.419773 PDT | [0] Epoch -172 finished +---------------------------------- --------------- +epoch -172 +replay_buffer/size 999033 +trainer/num train calls 829000 +trainer/Policy Loss -21.1672 +trainer/Log Pis Mean 25.4064 +trainer/Log Pis Std 13.2371 +trainer/Log Pis Max 70.6345 +trainer/Log Pis Min -8.24687 +trainer/policy/mean Mean -0.0413931 +trainer/policy/mean Std 0.912685 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76166 +trainer/policy/normal/std Std 0.649031 +trainer/policy/normal/std Max 6.78559 +trainer/policy/normal/std Min 0.241667 +trainer/policy/normal/log_std Mean 0.974521 +trainer/policy/normal/log_std Std 0.329508 +trainer/policy/normal/log_std Max 1.9148 +trainer/policy/normal/log_std Min -1.42019 +eval/num steps total 825320 +eval/num paths total 829 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.352873 +eval/Actions Std 0.862426 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73981 +time/logging (s) 0.00400763 +time/sampling batch (s) 0.289732 +time/saving (s) 0.003645 +time/training (s) 8.26172 +time/epoch (s) 11.2989 +time/total (s) 8450.86 +Epoch -172 +---------------------------------- --------------- +2022-05-10 15:31:49.246987 PDT | [0] Epoch -171 finished +---------------------------------- --------------- +epoch -171 +replay_buffer/size 999033 +trainer/num train calls 830000 +trainer/Policy Loss -20.5785 +trainer/Log Pis Mean 25.1023 +trainer/Log Pis Std 13.6595 +trainer/Log Pis Max 68.7114 +trainer/Log Pis Min -6.47705 +trainer/policy/mean Mean -0.0443107 +trainer/policy/mean Std 0.906981 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.80941 +trainer/policy/normal/std Std 0.68852 +trainer/policy/normal/std Max 6.91368 +trainer/policy/normal/std Min 0.266101 +trainer/policy/normal/log_std Mean 0.987042 +trainer/policy/normal/log_std Std 0.348447 +trainer/policy/normal/log_std Max 1.9335 +trainer/policy/normal/log_std Min -1.32388 +eval/num steps total 826320 +eval/num paths total 830 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.142271 +eval/Actions Std 0.921702 +eval/Actions Max 0.999996 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56241 +time/logging (s) 0.00366755 +time/sampling batch (s) 0.283388 +time/saving (s) 0.00343673 +time/training (s) 6.94958 +time/epoch (s) 9.80247 +time/total (s) 8460.67 +Epoch -171 +---------------------------------- --------------- +2022-05-10 15:31:59.184669 PDT | [0] Epoch -170 finished +---------------------------------- --------------- +epoch -170 +replay_buffer/size 999033 +trainer/num train calls 831000 +trainer/Policy Loss -19.8539 +trainer/Log Pis Mean 23.7262 +trainer/Log Pis Std 13.4881 +trainer/Log Pis Max 62.6947 +trainer/Log Pis Min -6.83684 +trainer/policy/mean Mean -0.0370051 +trainer/policy/mean Std 0.90885 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81346 +trainer/policy/normal/std Std 0.646392 +trainer/policy/normal/std Max 6.53496 +trainer/policy/normal/std Min 0.293553 +trainer/policy/normal/log_std Mean 0.997574 +trainer/policy/normal/log_std Std 0.30364 +trainer/policy/normal/log_std Max 1.87717 +trainer/policy/normal/log_std Min -1.2257 +eval/num steps total 827320 +eval/num paths total 831 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0725651 +eval/Actions Std 0.904034 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.56826 +time/logging (s) 0.00457012 +time/sampling batch (s) 0.29576 +time/saving (s) 0.00471279 +time/training (s) 7.0413 +time/epoch (s) 9.91461 +time/total (s) 8470.58 +Epoch -170 +---------------------------------- --------------- +2022-05-10 15:32:09.105744 PDT | [0] Epoch -169 finished +---------------------------------- --------------- +epoch -169 +replay_buffer/size 999033 +trainer/num train calls 832000 +trainer/Policy Loss -21.4267 +trainer/Log Pis Mean 26.4668 +trainer/Log Pis Std 13.6818 +trainer/Log Pis Max 74.4435 +trainer/Log Pis Min -7.41524 +trainer/policy/mean Mean -0.0627797 +trainer/policy/mean Std 0.909871 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999982 +trainer/policy/normal/std Mean 2.81511 +trainer/policy/normal/std Std 0.666088 +trainer/policy/normal/std Max 5.71288 +trainer/policy/normal/std Min 0.29811 +trainer/policy/normal/log_std Mean 0.992562 +trainer/policy/normal/log_std Std 0.333359 +trainer/policy/normal/log_std Max 1.74272 +trainer/policy/normal/log_std Min -1.21029 +eval/num steps total 828320 +eval/num paths total 832 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.176978 +eval/Actions Std 0.802749 +eval/Actions Max 0.999977 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75448 +time/logging (s) 0.00424901 +time/sampling batch (s) 0.283532 +time/saving (s) 0.00445898 +time/training (s) 6.84782 +time/epoch (s) 9.89455 +time/total (s) 8480.48 +Epoch -169 +---------------------------------- --------------- +2022-05-10 15:32:18.829852 PDT | [0] Epoch -168 finished +---------------------------------- --------------- +epoch -168 +replay_buffer/size 999033 +trainer/num train calls 833000 +trainer/Policy Loss -19.9849 +trainer/Log Pis Mean 24.8546 +trainer/Log Pis Std 13.196 +trainer/Log Pis Max 77.4054 +trainer/Log Pis Min -9.96273 +trainer/policy/mean Mean -0.0364316 +trainer/policy/mean Std 0.909387 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.80775 +trainer/policy/normal/std Std 0.651188 +trainer/policy/normal/std Max 6.15407 +trainer/policy/normal/std Min 0.25598 +trainer/policy/normal/log_std Mean 0.99352 +trainer/policy/normal/log_std Std 0.315181 +trainer/policy/normal/log_std Max 1.81711 +trainer/policy/normal/log_std Min -1.36265 +eval/num steps total 829320 +eval/num paths total 833 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114655 +eval/Actions Std 0.917744 +eval/Actions Max 0.999994 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78909 +time/logging (s) 0.00378247 +time/sampling batch (s) 0.534586 +time/saving (s) 0.00353777 +time/training (s) 6.36799 +time/epoch (s) 9.69899 +time/total (s) 8490.19 +Epoch -168 +---------------------------------- --------------- +2022-05-10 15:32:28.732070 PDT | [0] Epoch -167 finished +---------------------------------- --------------- +epoch -167 +replay_buffer/size 999033 +trainer/num train calls 834000 +trainer/Policy Loss -21.1128 +trainer/Log Pis Mean 25.0469 +trainer/Log Pis Std 13.5221 +trainer/Log Pis Max 63.0892 +trainer/Log Pis Min -6.39176 +trainer/policy/mean Mean -0.0570261 +trainer/policy/mean Std 0.90922 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83447 +trainer/policy/normal/std Std 0.703697 +trainer/policy/normal/std Max 5.99841 +trainer/policy/normal/std Min 0.260892 +trainer/policy/normal/log_std Mean 0.99519 +trainer/policy/normal/log_std Std 0.351015 +trainer/policy/normal/log_std Max 1.79149 +trainer/policy/normal/log_std Min -1.34365 +eval/num steps total 830320 +eval/num paths total 834 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.476079 +eval/Actions Std 0.699358 +eval/Actions Max 0.999979 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79626 +time/logging (s) 0.00394378 +time/sampling batch (s) 0.282697 +time/saving (s) 0.00415569 +time/training (s) 6.79083 +time/epoch (s) 9.87789 +time/total (s) 8500.07 +Epoch -167 +---------------------------------- --------------- +2022-05-10 15:32:38.463398 PDT | [0] Epoch -166 finished +---------------------------------- --------------- +epoch -166 +replay_buffer/size 999033 +trainer/num train calls 835000 +trainer/Policy Loss -18.9356 +trainer/Log Pis Mean 25.7847 +trainer/Log Pis Std 13.9846 +trainer/Log Pis Max 93.6346 +trainer/Log Pis Min -5.84675 +trainer/policy/mean Mean -0.0420332 +trainer/policy/mean Std 0.901382 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.89555 +trainer/policy/normal/std Std 0.699933 +trainer/policy/normal/std Max 6.10659 +trainer/policy/normal/std Min 0.331752 +trainer/policy/normal/log_std Mean 1.02006 +trainer/policy/normal/log_std Std 0.3341 +trainer/policy/normal/log_std Max 1.80937 +trainer/policy/normal/log_std Min -1.10337 +eval/num steps total 831320 +eval/num paths total 835 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.393417 +eval/Actions Std 0.769355 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81269 +time/logging (s) 0.00410849 +time/sampling batch (s) 0.280853 +time/saving (s) 0.00417218 +time/training (s) 6.60591 +time/epoch (s) 9.70773 +time/total (s) 8509.78 +Epoch -166 +---------------------------------- --------------- +2022-05-10 15:32:48.372063 PDT | [0] Epoch -165 finished +---------------------------------- --------------- +epoch -165 +replay_buffer/size 999033 +trainer/num train calls 836000 +trainer/Policy Loss -20.7848 +trainer/Log Pis Mean 25.4694 +trainer/Log Pis Std 13.4658 +trainer/Log Pis Max 71.4624 +trainer/Log Pis Min -11.8954 +trainer/policy/mean Mean -0.0234219 +trainer/policy/mean Std 0.907834 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999978 +trainer/policy/normal/std Mean 2.81566 +trainer/policy/normal/std Std 0.669908 +trainer/policy/normal/std Max 6.15091 +trainer/policy/normal/std Min 0.281534 +trainer/policy/normal/log_std Mean 0.993799 +trainer/policy/normal/log_std Std 0.326452 +trainer/policy/normal/log_std Max 1.8166 +trainer/policy/normal/log_std Min -1.2675 +eval/num steps total 832320 +eval/num paths total 836 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0471521 +eval/Actions Std 0.902288 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81949 +time/logging (s) 0.00392209 +time/sampling batch (s) 0.282498 +time/saving (s) 0.00355466 +time/training (s) 6.77454 +time/epoch (s) 9.884 +time/total (s) 8519.67 +Epoch -165 +---------------------------------- --------------- +2022-05-10 15:32:59.075178 PDT | [0] Epoch -164 finished +---------------------------------- --------------- +epoch -164 +replay_buffer/size 999033 +trainer/num train calls 837000 +trainer/Policy Loss -19.9055 +trainer/Log Pis Mean 24.4571 +trainer/Log Pis Std 13.678 +trainer/Log Pis Max 71.3949 +trainer/Log Pis Min -16.6282 +trainer/policy/mean Mean -0.0568147 +trainer/policy/mean Std 0.901869 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.76311 +trainer/policy/normal/std Std 0.688115 +trainer/policy/normal/std Max 7.26879 +trainer/policy/normal/std Min 0.203867 +trainer/policy/normal/log_std Mean 0.971448 +trainer/policy/normal/log_std Std 0.340893 +trainer/policy/normal/log_std Max 1.98359 +trainer/policy/normal/log_std Min -1.59029 +eval/num steps total 833320 +eval/num paths total 837 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0194116 +eval/Actions Std 0.909981 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79516 +time/logging (s) 0.0038058 +time/sampling batch (s) 0.281247 +time/saving (s) 0.00347991 +time/training (s) 7.59572 +time/epoch (s) 10.6794 +time/total (s) 8530.35 +Epoch -164 +---------------------------------- --------------- +2022-05-10 15:33:09.407126 PDT | [0] Epoch -163 finished +---------------------------------- --------------- +epoch -163 +replay_buffer/size 999033 +trainer/num train calls 838000 +trainer/Policy Loss -20.2996 +trainer/Log Pis Mean 25.5198 +trainer/Log Pis Std 14.0801 +trainer/Log Pis Max 79.2563 +trainer/Log Pis Min -10.9768 +trainer/policy/mean Mean -0.0365566 +trainer/policy/mean Std 0.907075 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85058 +trainer/policy/normal/std Std 0.667999 +trainer/policy/normal/std Max 5.81166 +trainer/policy/normal/std Min 0.320055 +trainer/policy/normal/log_std Mean 1.00834 +trainer/policy/normal/log_std Std 0.316061 +trainer/policy/normal/log_std Max 1.75987 +trainer/policy/normal/log_std Min -1.13926 +eval/num steps total 834320 +eval/num paths total 838 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.227342 +eval/Actions Std 0.893218 +eval/Actions Max 1 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42614 +time/logging (s) 0.00379432 +time/sampling batch (s) 0.279742 +time/saving (s) 0.00343187 +time/training (s) 7.59524 +time/epoch (s) 10.3083 +time/total (s) 8540.66 +Epoch -163 +---------------------------------- --------------- +2022-05-10 15:33:19.477958 PDT | [0] Epoch -162 finished +---------------------------------- --------------- +epoch -162 +replay_buffer/size 999033 +trainer/num train calls 839000 +trainer/Policy Loss -20.1314 +trainer/Log Pis Mean 24.9659 +trainer/Log Pis Std 12.6318 +trainer/Log Pis Max 69.1089 +trainer/Log Pis Min -6.33265 +trainer/policy/mean Mean -0.0552658 +trainer/policy/mean Std 0.909606 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.84928 +trainer/policy/normal/std Std 0.672841 +trainer/policy/normal/std Max 6.31398 +trainer/policy/normal/std Min 0.263131 +trainer/policy/normal/log_std Mean 1.00673 +trainer/policy/normal/log_std Std 0.321712 +trainer/policy/normal/log_std Max 1.84277 +trainer/policy/normal/log_std Min -1.3351 +eval/num steps total 835320 +eval/num paths total 839 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.179175 +eval/Actions Std 0.862903 +eval/Actions Max 0.999977 +eval/Actions Min -0.999985 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69195 +time/logging (s) 0.00437766 +time/sampling batch (s) 0.287383 +time/saving (s) 0.00439333 +time/training (s) 7.05921 +time/epoch (s) 10.0473 +time/total (s) 8550.71 +Epoch -162 +---------------------------------- --------------- +2022-05-10 15:33:31.218516 PDT | [0] Epoch -161 finished +---------------------------------- --------------- +epoch -161 +replay_buffer/size 999033 +trainer/num train calls 840000 +trainer/Policy Loss -19.6588 +trainer/Log Pis Mean 24.7573 +trainer/Log Pis Std 13.9741 +trainer/Log Pis Max 68.0822 +trainer/Log Pis Min -7.70595 +trainer/policy/mean Mean -0.00222825 +trainer/policy/mean Std 0.902524 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80046 +trainer/policy/normal/std Std 0.679571 +trainer/policy/normal/std Max 5.60395 +trainer/policy/normal/std Min 0.282275 +trainer/policy/normal/log_std Mean 0.987195 +trainer/policy/normal/log_std Std 0.329431 +trainer/policy/normal/log_std Max 1.72347 +trainer/policy/normal/log_std Min -1.26487 +eval/num steps total 836320 +eval/num paths total 840 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.1427 +eval/Actions Std 0.819912 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72316 +time/logging (s) 0.00379983 +time/sampling batch (s) 0.541362 +time/saving (s) 0.00363619 +time/training (s) 8.44294 +time/epoch (s) 11.7149 +time/total (s) 8562.43 +Epoch -161 +---------------------------------- --------------- +2022-05-10 15:33:42.180306 PDT | [0] Epoch -160 finished +---------------------------------- --------------- +epoch -160 +replay_buffer/size 999033 +trainer/num train calls 841000 +trainer/Policy Loss -20.4788 +trainer/Log Pis Mean 26.0598 +trainer/Log Pis Std 14.4119 +trainer/Log Pis Max 79.1492 +trainer/Log Pis Min -5.74502 +trainer/policy/mean Mean -0.022599 +trainer/policy/mean Std 0.905671 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.85654 +trainer/policy/normal/std Std 0.675173 +trainer/policy/normal/std Max 5.23046 +trainer/policy/normal/std Min 0.257837 +trainer/policy/normal/log_std Mean 1.00892 +trainer/policy/normal/log_std Std 0.323306 +trainer/policy/normal/log_std Max 1.6545 +trainer/policy/normal/log_std Min -1.35543 +eval/num steps total 837320 +eval/num paths total 841 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.133743 +eval/Actions Std 0.88687 +eval/Actions Max 0.999994 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66537 +time/logging (s) 0.00380221 +time/sampling batch (s) 0.533871 +time/saving (s) 0.00347296 +time/training (s) 7.73152 +time/epoch (s) 10.938 +time/total (s) 8573.37 +Epoch -160 +---------------------------------- --------------- +2022-05-10 15:33:51.578912 PDT | [0] Epoch -159 finished +---------------------------------- --------------- +epoch -159 +replay_buffer/size 999033 +trainer/num train calls 842000 +trainer/Policy Loss -20.6862 +trainer/Log Pis Mean 23.4662 +trainer/Log Pis Std 12.9517 +trainer/Log Pis Max 61.3187 +trainer/Log Pis Min -7.12812 +trainer/policy/mean Mean -0.043242 +trainer/policy/mean Std 0.908967 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79921 +trainer/policy/normal/std Std 0.652931 +trainer/policy/normal/std Max 6.69518 +trainer/policy/normal/std Min 0.348965 +trainer/policy/normal/log_std Mean 0.990992 +trainer/policy/normal/log_std Std 0.31084 +trainer/policy/normal/log_std Max 1.90139 +trainer/policy/normal/log_std Min -1.05278 +eval/num steps total 838320 +eval/num paths total 842 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.245646 +eval/Actions Std 0.816109 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.4666 +time/logging (s) 0.00373797 +time/sampling batch (s) 0.280759 +time/saving (s) 0.00346222 +time/training (s) 6.62039 +time/epoch (s) 9.37494 +time/total (s) 8582.75 +Epoch -159 +---------------------------------- --------------- +2022-05-10 15:34:01.297238 PDT | [0] Epoch -158 finished +---------------------------------- --------------- +epoch -158 +replay_buffer/size 999033 +trainer/num train calls 843000 +trainer/Policy Loss -19.7248 +trainer/Log Pis Mean 24.156 +trainer/Log Pis Std 12.6425 +trainer/Log Pis Max 72.3582 +trainer/Log Pis Min -8.4859 +trainer/policy/mean Mean -0.0353633 +trainer/policy/mean Std 0.90716 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.7327 +trainer/policy/normal/std Std 0.660283 +trainer/policy/normal/std Max 5.47333 +trainer/policy/normal/std Min 0.280967 +trainer/policy/normal/log_std Mean 0.962095 +trainer/policy/normal/log_std Std 0.334392 +trainer/policy/normal/log_std Max 1.69989 +trainer/policy/normal/log_std Min -1.26952 +eval/num steps total 839320 +eval/num paths total 843 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.11375 +eval/Actions Std 0.929043 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49546 +time/logging (s) 0.0038531 +time/sampling batch (s) 0.28091 +time/saving (s) 0.00346346 +time/training (s) 6.91103 +time/epoch (s) 9.69471 +time/total (s) 8592.44 +Epoch -158 +---------------------------------- --------------- +2022-05-10 15:34:10.463035 PDT | [0] Epoch -157 finished +---------------------------------- --------------- +epoch -157 +replay_buffer/size 999033 +trainer/num train calls 844000 +trainer/Policy Loss -19.6457 +trainer/Log Pis Mean 24.2524 +trainer/Log Pis Std 12.9895 +trainer/Log Pis Max 67.7679 +trainer/Log Pis Min -8.12673 +trainer/policy/mean Mean -0.0333881 +trainer/policy/mean Std 0.904358 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.81575 +trainer/policy/normal/std Std 0.649398 +trainer/policy/normal/std Max 5.41814 +trainer/policy/normal/std Min 0.328496 +trainer/policy/normal/log_std Mean 0.998271 +trainer/policy/normal/log_std Std 0.303711 +trainer/policy/normal/log_std Max 1.68975 +trainer/policy/normal/log_std Min -1.11323 +eval/num steps total 840320 +eval/num paths total 844 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0507944 +eval/Actions Std 0.899773 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.38836 +time/logging (s) 0.00372181 +time/sampling batch (s) 0.280168 +time/saving (s) 0.00394872 +time/training (s) 6.46579 +time/epoch (s) 9.142 +time/total (s) 8601.59 +Epoch -157 +---------------------------------- --------------- +2022-05-10 15:34:19.210603 PDT | [0] Epoch -156 finished +---------------------------------- --------------- +epoch -156 +replay_buffer/size 999033 +trainer/num train calls 845000 +trainer/Policy Loss -18.5097 +trainer/Log Pis Mean 24.1729 +trainer/Log Pis Std 13.2362 +trainer/Log Pis Max 74.975 +trainer/Log Pis Min -7.9374 +trainer/policy/mean Mean -0.0324515 +trainer/policy/mean Std 0.907037 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.80073 +trainer/policy/normal/std Std 0.642543 +trainer/policy/normal/std Max 5.53822 +trainer/policy/normal/std Min 0.315221 +trainer/policy/normal/log_std Mean 0.992577 +trainer/policy/normal/log_std Std 0.307111 +trainer/policy/normal/log_std Max 1.71167 +trainer/policy/normal/log_std Min -1.15448 +eval/num steps total 841320 +eval/num paths total 845 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0559756 +eval/Actions Std 0.922343 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42315 +time/logging (s) 0.00369606 +time/sampling batch (s) 0.281247 +time/saving (s) 0.00350854 +time/training (s) 6.01145 +time/epoch (s) 8.72306 +time/total (s) 8610.31 +Epoch -156 +---------------------------------- --------------- +2022-05-10 15:34:28.151220 PDT | [0] Epoch -155 finished +---------------------------------- --------------- +epoch -155 +replay_buffer/size 999033 +trainer/num train calls 846000 +trainer/Policy Loss -19.0345 +trainer/Log Pis Mean 24.2483 +trainer/Log Pis Std 13.3277 +trainer/Log Pis Max 67.1977 +trainer/Log Pis Min -9.56382 +trainer/policy/mean Mean -0.0514422 +trainer/policy/mean Std 0.904912 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.78379 +trainer/policy/normal/std Std 0.679655 +trainer/policy/normal/std Max 6.23752 +trainer/policy/normal/std Min 0.258559 +trainer/policy/normal/log_std Mean 0.978083 +trainer/policy/normal/log_std Std 0.348206 +trainer/policy/normal/log_std Max 1.83058 +trainer/policy/normal/log_std Min -1.35263 +eval/num steps total 842320 +eval/num paths total 846 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101538 +eval/Actions Std 0.919129 +eval/Actions Max 0.999992 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.37848 +time/logging (s) 0.00373841 +time/sampling batch (s) 0.281719 +time/saving (s) 0.00347955 +time/training (s) 6.24963 +time/epoch (s) 8.91705 +time/total (s) 8619.23 +Epoch -155 +---------------------------------- --------------- +2022-05-10 15:34:37.439323 PDT | [0] Epoch -154 finished +---------------------------------- --------------- +epoch -154 +replay_buffer/size 999033 +trainer/num train calls 847000 +trainer/Policy Loss -20.2167 +trainer/Log Pis Mean 25.178 +trainer/Log Pis Std 13.3485 +trainer/Log Pis Max 67.2855 +trainer/Log Pis Min -8.68778 +trainer/policy/mean Mean -0.0469129 +trainer/policy/mean Std 0.908225 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.79637 +trainer/policy/normal/std Std 0.666198 +trainer/policy/normal/std Max 6.14205 +trainer/policy/normal/std Min 0.297124 +trainer/policy/normal/log_std Mean 0.987389 +trainer/policy/normal/log_std Std 0.322577 +trainer/policy/normal/log_std Max 1.81516 +trainer/policy/normal/log_std Min -1.2136 +eval/num steps total 843320 +eval/num paths total 847 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.11951 +eval/Actions Std 0.944606 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41327 +time/logging (s) 0.00431582 +time/sampling batch (s) 0.292315 +time/saving (s) 0.00421139 +time/training (s) 6.55054 +time/epoch (s) 9.26466 +time/total (s) 8628.5 +Epoch -154 +---------------------------------- --------------- +2022-05-10 15:34:47.202723 PDT | [0] Epoch -153 finished +---------------------------------- --------------- +epoch -153 +replay_buffer/size 999033 +trainer/num train calls 848000 +trainer/Policy Loss -19.7115 +trainer/Log Pis Mean 23.8478 +trainer/Log Pis Std 12.8561 +trainer/Log Pis Max 62.7991 +trainer/Log Pis Min -10.8174 +trainer/policy/mean Mean -0.0399617 +trainer/policy/mean Std 0.90265 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78517 +trainer/policy/normal/std Std 0.673478 +trainer/policy/normal/std Max 6.90919 +trainer/policy/normal/std Min 0.253696 +trainer/policy/normal/log_std Mean 0.982685 +trainer/policy/normal/log_std Std 0.325197 +trainer/policy/normal/log_std Max 1.93285 +trainer/policy/normal/log_std Min -1.37162 +eval/num steps total 844320 +eval/num paths total 848 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.331152 +eval/Actions Std 0.840149 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48627 +time/logging (s) 0.00423368 +time/sampling batch (s) 0.291245 +time/saving (s) 0.00432425 +time/training (s) 6.95201 +time/epoch (s) 9.73808 +time/total (s) 8638.24 +Epoch -153 +---------------------------------- --------------- +2022-05-10 15:34:58.210219 PDT | [0] Epoch -152 finished +---------------------------------- --------------- +epoch -152 +replay_buffer/size 999033 +trainer/num train calls 849000 +trainer/Policy Loss -19.8094 +trainer/Log Pis Mean 24.6321 +trainer/Log Pis Std 13.4148 +trainer/Log Pis Max 63.3476 +trainer/Log Pis Min -9.09487 +trainer/policy/mean Mean -0.0466865 +trainer/policy/mean Std 0.907291 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81293 +trainer/policy/normal/std Std 0.681067 +trainer/policy/normal/std Max 6.0352 +trainer/policy/normal/std Min 0.349058 +trainer/policy/normal/log_std Mean 0.993978 +trainer/policy/normal/log_std Std 0.31561 +trainer/policy/normal/log_std Max 1.79761 +trainer/policy/normal/log_std Min -1.05252 +eval/num steps total 845320 +eval/num paths total 849 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.238225 +eval/Actions Std 0.944932 +eval/Actions Max 0.999987 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63321 +time/logging (s) 0.00510355 +time/sampling batch (s) 0.544571 +time/saving (s) 0.00486838 +time/training (s) 7.7953 +time/epoch (s) 10.9831 +time/total (s) 8649.23 +Epoch -152 +---------------------------------- --------------- +2022-05-10 15:35:09.160407 PDT | [0] Epoch -151 finished +---------------------------------- --------------- +epoch -151 +replay_buffer/size 999033 +trainer/num train calls 850000 +trainer/Policy Loss -18.7837 +trainer/Log Pis Mean 24.2479 +trainer/Log Pis Std 13.3672 +trainer/Log Pis Max 67.4085 +trainer/Log Pis Min -6.47201 +trainer/policy/mean Mean -0.0521883 +trainer/policy/mean Std 0.904645 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.7547 +trainer/policy/normal/std Std 0.669093 +trainer/policy/normal/std Max 5.42699 +trainer/policy/normal/std Min 0.308479 +trainer/policy/normal/log_std Mean 0.969147 +trainer/policy/normal/log_std Std 0.338406 +trainer/policy/normal/log_std Max 1.69138 +trainer/policy/normal/log_std Min -1.1761 +eval/num steps total 846320 +eval/num paths total 850 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00698857 +eval/Actions Std 0.888579 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59024 +time/logging (s) 0.00417731 +time/sampling batch (s) 0.791912 +time/saving (s) 0.00367328 +time/training (s) 7.53259 +time/epoch (s) 10.9226 +time/total (s) 8660.16 +Epoch -151 +---------------------------------- --------------- +2022-05-10 15:35:19.018839 PDT | [0] Epoch -150 finished +---------------------------------- --------------- +epoch -150 +replay_buffer/size 999033 +trainer/num train calls 851000 +trainer/Policy Loss -19.9658 +trainer/Log Pis Mean 24.2371 +trainer/Log Pis Std 13.2234 +trainer/Log Pis Max 68.6659 +trainer/Log Pis Min -8.79092 +trainer/policy/mean Mean -0.0485258 +trainer/policy/mean Std 0.903816 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.8005 +trainer/policy/normal/std Std 0.692331 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.290797 +trainer/policy/normal/log_std Mean 0.984741 +trainer/policy/normal/log_std Std 0.341223 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.23513 +eval/num steps total 847320 +eval/num paths total 851 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.326751 +eval/Actions Std 0.840912 +eval/Actions Max 0.999996 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63807 +time/logging (s) 0.00408997 +time/sampling batch (s) 0.291382 +time/saving (s) 0.00352538 +time/training (s) 6.89647 +time/epoch (s) 9.83354 +time/total (s) 8669.99 +Epoch -150 +---------------------------------- --------------- +2022-05-10 15:35:28.767781 PDT | [0] Epoch -149 finished +---------------------------------- --------------- +epoch -149 +replay_buffer/size 999033 +trainer/num train calls 852000 +trainer/Policy Loss -18.9734 +trainer/Log Pis Mean 24.0697 +trainer/Log Pis Std 14.005 +trainer/Log Pis Max 78.6677 +trainer/Log Pis Min -14.0239 +trainer/policy/mean Mean -0.0420585 +trainer/policy/mean Std 0.908153 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79132 +trainer/policy/normal/std Std 0.66549 +trainer/policy/normal/std Max 5.33703 +trainer/policy/normal/std Min 0.298574 +trainer/policy/normal/log_std Mean 0.984078 +trainer/policy/normal/log_std Std 0.332476 +trainer/policy/normal/log_std Max 1.67467 +trainer/policy/normal/log_std Min -1.20874 +eval/num steps total 848320 +eval/num paths total 852 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0210892 +eval/Actions Std 0.948906 +eval/Actions Max 1 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6558 +time/logging (s) 0.00417306 +time/sampling batch (s) 0.292057 +time/saving (s) 0.00421171 +time/training (s) 6.76814 +time/epoch (s) 9.72438 +time/total (s) 8679.72 +Epoch -149 +---------------------------------- --------------- +2022-05-10 15:35:38.829021 PDT | [0] Epoch -148 finished +---------------------------------- --------------- +epoch -148 +replay_buffer/size 999033 +trainer/num train calls 853000 +trainer/Policy Loss -19.6783 +trainer/Log Pis Mean 25.1934 +trainer/Log Pis Std 13.3728 +trainer/Log Pis Max 65.044 +trainer/Log Pis Min -11.9645 +trainer/policy/mean Mean -0.0345328 +trainer/policy/mean Std 0.903942 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77019 +trainer/policy/normal/std Std 0.655497 +trainer/policy/normal/std Max 5.77163 +trainer/policy/normal/std Min 0.262644 +trainer/policy/normal/log_std Mean 0.978815 +trainer/policy/normal/log_std Std 0.319807 +trainer/policy/normal/log_std Max 1.75295 +trainer/policy/normal/log_std Min -1.33696 +eval/num steps total 849320 +eval/num paths total 853 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.101423 +eval/Actions Std 0.915645 +eval/Actions Max 0.999993 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67341 +time/logging (s) 0.00406553 +time/sampling batch (s) 0.292391 +time/saving (s) 0.00350961 +time/training (s) 7.06286 +time/epoch (s) 10.0362 +time/total (s) 8689.76 +Epoch -148 +---------------------------------- --------------- +2022-05-10 15:35:48.792059 PDT | [0] Epoch -147 finished +---------------------------------- --------------- +epoch -147 +replay_buffer/size 999033 +trainer/num train calls 854000 +trainer/Policy Loss -20.9101 +trainer/Log Pis Mean 24.0059 +trainer/Log Pis Std 12.3586 +trainer/Log Pis Max 62.0205 +trainer/Log Pis Min -11.0623 +trainer/policy/mean Mean -0.0376973 +trainer/policy/mean Std 0.91025 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.75512 +trainer/policy/normal/std Std 0.661399 +trainer/policy/normal/std Max 6.17646 +trainer/policy/normal/std Min 0.330463 +trainer/policy/normal/log_std Mean 0.972189 +trainer/policy/normal/log_std Std 0.323368 +trainer/policy/normal/log_std Max 1.82074 +trainer/policy/normal/log_std Min -1.10726 +eval/num steps total 850320 +eval/num paths total 854 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118705 +eval/Actions Std 0.918005 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81452 +time/logging (s) 0.00400766 +time/sampling batch (s) 0.291424 +time/saving (s) 0.0035246 +time/training (s) 6.82498 +time/epoch (s) 9.93845 +time/total (s) 8699.7 +Epoch -147 +---------------------------------- --------------- +2022-05-10 15:35:58.400292 PDT | [0] Epoch -146 finished +---------------------------------- --------------- +epoch -146 +replay_buffer/size 999033 +trainer/num train calls 855000 +trainer/Policy Loss -20.3275 +trainer/Log Pis Mean 25.3 +trainer/Log Pis Std 13.4315 +trainer/Log Pis Max 77.0082 +trainer/Log Pis Min -3.89382 +trainer/policy/mean Mean -0.044341 +trainer/policy/mean Std 0.911277 +trainer/policy/mean Max 1 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80533 +trainer/policy/normal/std Std 0.681097 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.297792 +trainer/policy/normal/log_std Mean 0.987063 +trainer/policy/normal/log_std Std 0.342014 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.21136 +eval/num steps total 851320 +eval/num paths total 855 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0880736 +eval/Actions Std 0.902485 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.59559 +time/logging (s) 0.00367456 +time/sampling batch (s) 0.282262 +time/saving (s) 0.00343443 +time/training (s) 6.69903 +time/epoch (s) 9.584 +time/total (s) 8709.29 +Epoch -146 +---------------------------------- --------------- +2022-05-10 15:36:08.047919 PDT | [0] Epoch -145 finished +---------------------------------- --------------- +epoch -145 +replay_buffer/size 999033 +trainer/num train calls 856000 +trainer/Policy Loss -20.3224 +trainer/Log Pis Mean 25.152 +trainer/Log Pis Std 13.5556 +trainer/Log Pis Max 68.9543 +trainer/Log Pis Min -6.64568 +trainer/policy/mean Mean -0.0291517 +trainer/policy/mean Std 0.908375 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83053 +trainer/policy/normal/std Std 0.670524 +trainer/policy/normal/std Max 6.07533 +trainer/policy/normal/std Min 0.264493 +trainer/policy/normal/log_std Mean 0.999163 +trainer/policy/normal/log_std Std 0.327529 +trainer/policy/normal/log_std Max 1.80424 +trainer/policy/normal/log_std Min -1.32994 +eval/num steps total 852320 +eval/num paths total 856 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.311959 +eval/Actions Std 0.868996 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.81096 +time/logging (s) 0.00400137 +time/sampling batch (s) 0.28163 +time/saving (s) 0.00417135 +time/training (s) 6.52334 +time/epoch (s) 9.6241 +time/total (s) 8718.91 +Epoch -145 +---------------------------------- --------------- +2022-05-10 15:36:17.657649 PDT | [0] Epoch -144 finished +---------------------------------- --------------- +epoch -144 +replay_buffer/size 999033 +trainer/num train calls 857000 +trainer/Policy Loss -19.7991 +trainer/Log Pis Mean 23.3243 +trainer/Log Pis Std 13.0264 +trainer/Log Pis Max 70.4808 +trainer/Log Pis Min -8.75146 +trainer/policy/mean Mean -0.0332262 +trainer/policy/mean Std 0.903779 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.81101 +trainer/policy/normal/std Std 0.664566 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.306223 +trainer/policy/normal/log_std Mean 0.993195 +trainer/policy/normal/log_std Std 0.321682 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.18344 +eval/num steps total 853320 +eval/num paths total 857 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118379 +eval/Actions Std 0.919419 +eval/Actions Max 1 +eval/Actions Min -0.999959 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60881 +time/logging (s) 0.00368386 +time/sampling batch (s) 0.282145 +time/saving (s) 0.00350634 +time/training (s) 6.68708 +time/epoch (s) 9.58523 +time/total (s) 8728.5 +Epoch -144 +---------------------------------- --------------- +2022-05-10 15:36:26.961311 PDT | [0] Epoch -143 finished +---------------------------------- --------------- +epoch -143 +replay_buffer/size 999033 +trainer/num train calls 858000 +trainer/Policy Loss -20.9832 +trainer/Log Pis Mean 24.3593 +trainer/Log Pis Std 13.4228 +trainer/Log Pis Max 81.386 +trainer/Log Pis Min -5.19173 +trainer/policy/mean Mean -0.0334104 +trainer/policy/mean Std 0.907343 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.79551 +trainer/policy/normal/std Std 0.686467 +trainer/policy/normal/std Max 5.99753 +trainer/policy/normal/std Min 0.258926 +trainer/policy/normal/log_std Mean 0.985135 +trainer/policy/normal/log_std Std 0.329717 +trainer/policy/normal/log_std Max 1.79135 +trainer/policy/normal/log_std Min -1.35121 +eval/num steps total 854320 +eval/num paths total 858 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0363828 +eval/Actions Std 0.905336 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65745 +time/logging (s) 0.00365498 +time/sampling batch (s) 0.282062 +time/saving (s) 0.00347388 +time/training (s) 6.33312 +time/epoch (s) 9.27976 +time/total (s) 8737.79 +Epoch -143 +---------------------------------- --------------- +2022-05-10 15:36:37.134708 PDT | [0] Epoch -142 finished +---------------------------------- --------------- +epoch -142 +replay_buffer/size 999033 +trainer/num train calls 859000 +trainer/Policy Loss -19.7515 +trainer/Log Pis Mean 23.8147 +trainer/Log Pis Std 13.6332 +trainer/Log Pis Max 84.4128 +trainer/Log Pis Min -7.4215 +trainer/policy/mean Mean -0.0184936 +trainer/policy/mean Std 0.906846 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82172 +trainer/policy/normal/std Std 0.665731 +trainer/policy/normal/std Max 6.09648 +trainer/policy/normal/std Min 0.288148 +trainer/policy/normal/log_std Mean 0.997137 +trainer/policy/normal/log_std Std 0.321144 +trainer/policy/normal/log_std Max 1.80771 +trainer/policy/normal/log_std Min -1.24428 +eval/num steps total 855320 +eval/num paths total 859 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0017099 +eval/Actions Std 0.906752 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.41755 +time/logging (s) 0.0037023 +time/sampling batch (s) 1.03269 +time/saving (s) 0.00349297 +time/training (s) 6.69219 +time/epoch (s) 10.1496 +time/total (s) 8747.94 +Epoch -142 +---------------------------------- --------------- +2022-05-10 15:36:47.054917 PDT | [0] Epoch -141 finished +---------------------------------- --------------- +epoch -141 +replay_buffer/size 999033 +trainer/num train calls 860000 +trainer/Policy Loss -20.227 +trainer/Log Pis Mean 24.2531 +trainer/Log Pis Std 14.0889 +trainer/Log Pis Max 70.6413 +trainer/Log Pis Min -10.1994 +trainer/policy/mean Mean -0.021296 +trainer/policy/mean Std 0.906322 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76437 +trainer/policy/normal/std Std 0.665654 +trainer/policy/normal/std Max 6.61224 +trainer/policy/normal/std Min 0.344548 +trainer/policy/normal/log_std Mean 0.974748 +trainer/policy/normal/log_std Std 0.328529 +trainer/policy/normal/log_std Max 1.88892 +trainer/policy/normal/log_std Min -1.06552 +eval/num steps total 856320 +eval/num paths total 860 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00738791 +eval/Actions Std 0.913872 +eval/Actions Max 0.999999 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.80441 +time/logging (s) 0.00367954 +time/sampling batch (s) 0.283736 +time/saving (s) 0.00345925 +time/training (s) 6.80111 +time/epoch (s) 9.89639 +time/total (s) 8757.84 +Epoch -141 +---------------------------------- --------------- +2022-05-10 15:36:55.986727 PDT | [0] Epoch -140 finished +---------------------------------- --------------- +epoch -140 +replay_buffer/size 999033 +trainer/num train calls 861000 +trainer/Policy Loss -21.0274 +trainer/Log Pis Mean 24.3879 +trainer/Log Pis Std 12.8868 +trainer/Log Pis Max 61.4592 +trainer/Log Pis Min -8.57401 +trainer/policy/mean Mean -0.0305741 +trainer/policy/mean Std 0.910442 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.75713 +trainer/policy/normal/std Std 0.637063 +trainer/policy/normal/std Max 5.44425 +trainer/policy/normal/std Min 0.215207 +trainer/policy/normal/log_std Mean 0.975252 +trainer/policy/normal/log_std Std 0.315965 +trainer/policy/normal/log_std Max 1.69456 +trainer/policy/normal/log_std Min -1.53615 +eval/num steps total 857320 +eval/num paths total 861 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132018 +eval/Actions Std 0.910936 +eval/Actions Max 0.999944 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.65022 +time/logging (s) 0.0040239 +time/sampling batch (s) 0.283127 +time/saving (s) 0.00410488 +time/training (s) 5.96667 +time/epoch (s) 8.90814 +time/total (s) 8766.75 +Epoch -140 +---------------------------------- --------------- +2022-05-10 15:37:05.937233 PDT | [0] Epoch -139 finished +---------------------------------- --------------- +epoch -139 +replay_buffer/size 999033 +trainer/num train calls 862000 +trainer/Policy Loss -19.2237 +trainer/Log Pis Mean 24.1184 +trainer/Log Pis Std 13.2122 +trainer/Log Pis Max 62.1499 +trainer/Log Pis Min -12.2526 +trainer/policy/mean Mean -0.0293083 +trainer/policy/mean Std 0.907616 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.75565 +trainer/policy/normal/std Std 0.649057 +trainer/policy/normal/std Max 6.3196 +trainer/policy/normal/std Min 0.230763 +trainer/policy/normal/log_std Mean 0.974041 +trainer/policy/normal/log_std Std 0.316683 +trainer/policy/normal/log_std Max 1.84366 +trainer/policy/normal/log_std Min -1.46636 +eval/num steps total 858320 +eval/num paths total 862 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0242133 +eval/Actions Std 0.912365 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73484 +time/logging (s) 0.0037455 +time/sampling batch (s) 0.532963 +time/saving (s) 0.00357897 +time/training (s) 6.65087 +time/epoch (s) 9.926 +time/total (s) 8776.68 +Epoch -139 +---------------------------------- --------------- +2022-05-10 15:37:15.873271 PDT | [0] Epoch -138 finished +---------------------------------- --------------- +epoch -138 +replay_buffer/size 999033 +trainer/num train calls 863000 +trainer/Policy Loss -19.1993 +trainer/Log Pis Mean 24.0869 +trainer/Log Pis Std 13.8818 +trainer/Log Pis Max 63.6721 +trainer/Log Pis Min -6.75853 +trainer/policy/mean Mean -0.023234 +trainer/policy/mean Std 0.906732 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.7987 +trainer/policy/normal/std Std 0.696931 +trainer/policy/normal/std Max 6.55774 +trainer/policy/normal/std Min 0.25207 +trainer/policy/normal/log_std Mean 0.982341 +trainer/policy/normal/log_std Std 0.35075 +trainer/policy/normal/log_std Max 1.88065 +trainer/policy/normal/log_std Min -1.37805 +eval/num steps total 859320 +eval/num paths total 863 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.155356 +eval/Actions Std 0.864946 +eval/Actions Max 0.99998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70434 +time/logging (s) 0.00371296 +time/sampling batch (s) 0.281612 +time/saving (s) 0.00340167 +time/training (s) 6.91904 +time/epoch (s) 9.9121 +time/total (s) 8786.59 +Epoch -138 +---------------------------------- --------------- +2022-05-10 15:37:25.676046 PDT | [0] Epoch -137 finished +---------------------------------- --------------- +epoch -137 +replay_buffer/size 999033 +trainer/num train calls 864000 +trainer/Policy Loss -19.4077 +trainer/Log Pis Mean 24.8361 +trainer/Log Pis Std 13.6592 +trainer/Log Pis Max 75.0065 +trainer/Log Pis Min -6.03435 +trainer/policy/mean Mean -0.0477381 +trainer/policy/mean Std 0.90538 +trainer/policy/mean Max 0.999979 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.78912 +trainer/policy/normal/std Std 0.681252 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.217298 +trainer/policy/normal/log_std Mean 0.982952 +trainer/policy/normal/log_std Std 0.33057 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.52649 +eval/num steps total 860320 +eval/num paths total 864 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0544351 +eval/Actions Std 0.907355 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72411 +time/logging (s) 0.00366645 +time/sampling batch (s) 0.284194 +time/saving (s) 0.00343609 +time/training (s) 6.7634 +time/epoch (s) 9.77881 +time/total (s) 8796.37 +Epoch -137 +---------------------------------- --------------- +2022-05-10 15:37:35.701504 PDT | [0] Epoch -136 finished +---------------------------------- --------------- +epoch -136 +replay_buffer/size 999033 +trainer/num train calls 865000 +trainer/Policy Loss -19.5633 +trainer/Log Pis Mean 25.1547 +trainer/Log Pis Std 13.1422 +trainer/Log Pis Max 61.9236 +trainer/Log Pis Min -12.3762 +trainer/policy/mean Mean -0.0261139 +trainer/policy/mean Std 0.910544 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.75719 +trainer/policy/normal/std Std 0.667351 +trainer/policy/normal/std Max 5.62924 +trainer/policy/normal/std Min 0.245613 +trainer/policy/normal/log_std Mean 0.97041 +trainer/policy/normal/log_std Std 0.338006 +trainer/policy/normal/log_std Max 1.72797 +trainer/policy/normal/log_std Min -1.404 +eval/num steps total 861320 +eval/num paths total 865 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.178613 +eval/Actions Std 0.938057 +eval/Actions Max 0.999996 +eval/Actions Min -0.999987 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79819 +time/logging (s) 0.00369974 +time/sampling batch (s) 0.280252 +time/saving (s) 0.00339022 +time/training (s) 6.91602 +time/epoch (s) 10.0016 +time/total (s) 8806.38 +Epoch -136 +---------------------------------- --------------- +2022-05-10 15:37:45.934925 PDT | [0] Epoch -135 finished +---------------------------------- --------------- +epoch -135 +replay_buffer/size 999033 +trainer/num train calls 866000 +trainer/Policy Loss -20.7462 +trainer/Log Pis Mean 25.7247 +trainer/Log Pis Std 13.4442 +trainer/Log Pis Max 80.8392 +trainer/Log Pis Min -7.19984 +trainer/policy/mean Mean -0.0401447 +trainer/policy/mean Std 0.91091 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.8257 +trainer/policy/normal/std Std 0.660073 +trainer/policy/normal/std Max 5.90803 +trainer/policy/normal/std Min 0.257541 +trainer/policy/normal/log_std Mean 0.998968 +trainer/policy/normal/log_std Std 0.319269 +trainer/policy/normal/log_std Max 1.77631 +trainer/policy/normal/log_std Min -1.35658 +eval/num steps total 862320 +eval/num paths total 866 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0967747 +eval/Actions Std 0.85389 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55575 +time/logging (s) 0.00375648 +time/sampling batch (s) 0.532035 +time/saving (s) 0.00346327 +time/training (s) 7.1146 +time/epoch (s) 10.2096 +time/total (s) 8816.59 +Epoch -135 +---------------------------------- --------------- +2022-05-10 15:37:55.771133 PDT | [0] Epoch -134 finished +---------------------------------- --------------- +epoch -134 +replay_buffer/size 999033 +trainer/num train calls 867000 +trainer/Policy Loss -19.1532 +trainer/Log Pis Mean 24.8474 +trainer/Log Pis Std 12.8782 +trainer/Log Pis Max 71.7797 +trainer/Log Pis Min -4.98721 +trainer/policy/mean Mean -0.0347116 +trainer/policy/mean Std 0.903623 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.82358 +trainer/policy/normal/std Std 0.691277 +trainer/policy/normal/std Max 6.15195 +trainer/policy/normal/std Min 0.298166 +trainer/policy/normal/log_std Mean 0.99295 +trainer/policy/normal/log_std Std 0.342999 +trainer/policy/normal/log_std Max 1.81677 +trainer/policy/normal/log_std Min -1.21011 +eval/num steps total 863320 +eval/num paths total 867 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0287325 +eval/Actions Std 0.906143 +eval/Actions Max 1 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.79198 +time/logging (s) 0.0039608 +time/sampling batch (s) 0.28068 +time/saving (s) 0.00418968 +time/training (s) 6.73172 +time/epoch (s) 9.81253 +time/total (s) 8826.41 +Epoch -134 +---------------------------------- --------------- +2022-05-10 15:38:05.802234 PDT | [0] Epoch -133 finished +---------------------------------- --------------- +epoch -133 +replay_buffer/size 999033 +trainer/num train calls 868000 +trainer/Policy Loss -20.4006 +trainer/Log Pis Mean 24.4754 +trainer/Log Pis Std 13.0477 +trainer/Log Pis Max 62.7193 +trainer/Log Pis Min -8.28364 +trainer/policy/mean Mean -0.0285036 +trainer/policy/mean Std 0.906425 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82514 +trainer/policy/normal/std Std 0.682949 +trainer/policy/normal/std Max 5.85429 +trainer/policy/normal/std Min 0.294506 +trainer/policy/normal/log_std Mean 0.995784 +trainer/policy/normal/log_std Std 0.331684 +trainer/policy/normal/log_std Max 1.76717 +trainer/policy/normal/log_std Min -1.22245 +eval/num steps total 864320 +eval/num paths total 868 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107932 +eval/Actions Std 0.838055 +eval/Actions Max 0.999997 +eval/Actions Min -0.999986 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78901 +time/logging (s) 0.00382528 +time/sampling batch (s) 0.281718 +time/saving (s) 0.00351525 +time/training (s) 6.92864 +time/epoch (s) 10.0067 +time/total (s) 8836.42 +Epoch -133 +---------------------------------- --------------- +2022-05-10 15:38:16.517312 PDT | [0] Epoch -132 finished +---------------------------------- --------------- +epoch -132 +replay_buffer/size 999033 +trainer/num train calls 869000 +trainer/Policy Loss -18.8114 +trainer/Log Pis Mean 25.264 +trainer/Log Pis Std 14.9545 +trainer/Log Pis Max 78.5922 +trainer/Log Pis Min -5.14437 +trainer/policy/mean Mean -0.0300347 +trainer/policy/mean Std 0.902622 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.87908 +trainer/policy/normal/std Std 0.712671 +trainer/policy/normal/std Max 5.81428 +trainer/policy/normal/std Min 0.251941 +trainer/policy/normal/log_std Mean 1.01305 +trainer/policy/normal/log_std Std 0.337224 +trainer/policy/normal/log_std Max 1.76032 +trainer/policy/normal/log_std Min -1.37856 +eval/num steps total 865320 +eval/num paths total 869 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.230832 +eval/Actions Std 0.960985 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68051 +time/logging (s) 0.00376842 +time/sampling batch (s) 0.279802 +time/saving (s) 0.00344913 +time/training (s) 7.72393 +time/epoch (s) 10.6915 +time/total (s) 8847.11 +Epoch -132 +---------------------------------- --------------- +2022-05-10 15:38:26.428951 PDT | [0] Epoch -131 finished +---------------------------------- --------------- +epoch -131 +replay_buffer/size 999033 +trainer/num train calls 870000 +trainer/Policy Loss -21.0695 +trainer/Log Pis Mean 24.0558 +trainer/Log Pis Std 13.3245 +trainer/Log Pis Max 76.7751 +trainer/Log Pis Min -12.4647 +trainer/policy/mean Mean -0.0504231 +trainer/policy/mean Std 0.908266 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.8231 +trainer/policy/normal/std Std 0.6564 +trainer/policy/normal/std Max 6.40434 +trainer/policy/normal/std Min 0.28896 +trainer/policy/normal/log_std Mean 0.99842 +trainer/policy/normal/log_std Std 0.319027 +trainer/policy/normal/log_std Max 1.85698 +trainer/policy/normal/log_std Min -1.24147 +eval/num steps total 866320 +eval/num paths total 870 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.222374 +eval/Actions Std 0.891683 +eval/Actions Max 0.999996 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55453 +time/logging (s) 0.00370814 +time/sampling batch (s) 0.281027 +time/saving (s) 0.00347129 +time/training (s) 7.04509 +time/epoch (s) 9.88783 +time/total (s) 8857 +Epoch -131 +---------------------------------- --------------- +2022-05-10 15:38:36.140154 PDT | [0] Epoch -130 finished +---------------------------------- --------------- +epoch -130 +replay_buffer/size 999033 +trainer/num train calls 871000 +trainer/Policy Loss -19.5499 +trainer/Log Pis Mean 24.2743 +trainer/Log Pis Std 14.0012 +trainer/Log Pis Max 70.923 +trainer/Log Pis Min -9.25524 +trainer/policy/mean Mean -0.0430434 +trainer/policy/mean Std 0.906957 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76034 +trainer/policy/normal/std Std 0.679936 +trainer/policy/normal/std Max 5.90613 +trainer/policy/normal/std Min 0.267591 +trainer/policy/normal/log_std Mean 0.969316 +trainer/policy/normal/log_std Std 0.348257 +trainer/policy/normal/log_std Max 1.77599 +trainer/policy/normal/log_std Min -1.3183 +eval/num steps total 867320 +eval/num paths total 871 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00293903 +eval/Actions Std 0.909948 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66851 +time/logging (s) 0.00382898 +time/sampling batch (s) 0.283451 +time/saving (s) 0.00350321 +time/training (s) 6.72802 +time/epoch (s) 9.68732 +time/total (s) 8866.69 +Epoch -130 +---------------------------------- --------------- +2022-05-10 15:38:45.459313 PDT | [0] Epoch -129 finished +---------------------------------- --------------- +epoch -129 +replay_buffer/size 999033 +trainer/num train calls 872000 +trainer/Policy Loss -19.1107 +trainer/Log Pis Mean 24.4477 +trainer/Log Pis Std 12.8579 +trainer/Log Pis Max 65.1484 +trainer/Log Pis Min -5.90773 +trainer/policy/mean Mean -0.0295269 +trainer/policy/mean Std 0.903389 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.82073 +trainer/policy/normal/std Std 0.686978 +trainer/policy/normal/std Max 5.65688 +trainer/policy/normal/std Min 0.250906 +trainer/policy/normal/log_std Mean 0.993078 +trainer/policy/normal/log_std Std 0.338507 +trainer/policy/normal/log_std Max 1.73287 +trainer/policy/normal/log_std Min -1.38268 +eval/num steps total 868320 +eval/num paths total 872 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0126857 +eval/Actions Std 0.956926 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.87298 +time/logging (s) 0.00430939 +time/sampling batch (s) 0.287128 +time/saving (s) 0.00416792 +time/training (s) 6.12634 +time/epoch (s) 9.29492 +time/total (s) 8875.99 +Epoch -129 +---------------------------------- --------------- +2022-05-10 15:38:55.685221 PDT | [0] Epoch -128 finished +---------------------------------- --------------- +epoch -128 +replay_buffer/size 999033 +trainer/num train calls 873000 +trainer/Policy Loss -19.1315 +trainer/Log Pis Mean 24.487 +trainer/Log Pis Std 13.0136 +trainer/Log Pis Max 66.2738 +trainer/Log Pis Min -4.42806 +trainer/policy/mean Mean -0.0396656 +trainer/policy/mean Std 0.906358 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.99998 +trainer/policy/normal/std Mean 2.79939 +trainer/policy/normal/std Std 0.677635 +trainer/policy/normal/std Max 5.30225 +trainer/policy/normal/std Min 0.320614 +trainer/policy/normal/log_std Mean 0.986554 +trainer/policy/normal/log_std Std 0.331671 +trainer/policy/normal/log_std Max 1.66813 +trainer/policy/normal/log_std Min -1.13752 +eval/num steps total 869320 +eval/num paths total 873 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.087796 +eval/Actions Std 0.915806 +eval/Actions Max 0.999993 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84213 +time/logging (s) 0.00400626 +time/sampling batch (s) 0.292743 +time/saving (s) 0.00380776 +time/training (s) 7.05745 +time/epoch (s) 10.2001 +time/total (s) 8886.19 +Epoch -128 +---------------------------------- --------------- +2022-05-10 15:39:05.848935 PDT | [0] Epoch -127 finished +---------------------------------- --------------- +epoch -127 +replay_buffer/size 999033 +trainer/num train calls 874000 +trainer/Policy Loss -19.7396 +trainer/Log Pis Mean 23.5317 +trainer/Log Pis Std 13.2213 +trainer/Log Pis Max 59.8439 +trainer/Log Pis Min -7.15757 +trainer/policy/mean Mean -0.014748 +trainer/policy/mean Std 0.906055 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.74397 +trainer/policy/normal/std Std 0.651763 +trainer/policy/normal/std Max 5.44901 +trainer/policy/normal/std Min 0.28422 +trainer/policy/normal/log_std Mean 0.968289 +trainer/policy/normal/log_std Std 0.324231 +trainer/policy/normal/log_std Max 1.69543 +trainer/policy/normal/log_std Min -1.25801 +eval/num steps total 870320 +eval/num paths total 874 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.204226 +eval/Actions Std 0.885 +eval/Actions Max 0.999992 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.9184 +time/logging (s) 0.00377757 +time/sampling batch (s) 0.294084 +time/saving (s) 0.00355443 +time/training (s) 6.91841 +time/epoch (s) 10.1382 +time/total (s) 8896.33 +Epoch -127 +---------------------------------- --------------- +2022-05-10 15:39:15.724764 PDT | [0] Epoch -126 finished +---------------------------------- --------------- +epoch -126 +replay_buffer/size 999033 +trainer/num train calls 875000 +trainer/Policy Loss -20.6007 +trainer/Log Pis Mean 24.3761 +trainer/Log Pis Std 13.3888 +trainer/Log Pis Max 62.6549 +trainer/Log Pis Min -8.14974 +trainer/policy/mean Mean -0.0243918 +trainer/policy/mean Std 0.904276 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77235 +trainer/policy/normal/std Std 0.668398 +trainer/policy/normal/std Max 6.22262 +trainer/policy/normal/std Min 0.260413 +trainer/policy/normal/log_std Mean 0.976819 +trainer/policy/normal/log_std Std 0.333251 +trainer/policy/normal/log_std Max 1.82819 +trainer/policy/normal/log_std Min -1.34549 +eval/num steps total 871320 +eval/num paths total 875 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00725148 +eval/Actions Std 0.896543 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76245 +time/logging (s) 0.00392095 +time/sampling batch (s) 0.291172 +time/saving (s) 0.00360047 +time/training (s) 6.78994 +time/epoch (s) 9.85109 +time/total (s) 8906.19 +Epoch -126 +---------------------------------- --------------- +2022-05-10 15:39:25.137605 PDT | [0] Epoch -125 finished +---------------------------------- --------------- +epoch -125 +replay_buffer/size 999033 +trainer/num train calls 876000 +trainer/Policy Loss -20.7263 +trainer/Log Pis Mean 23.9621 +trainer/Log Pis Std 12.7608 +trainer/Log Pis Max 65.3926 +trainer/Log Pis Min -9.57419 +trainer/policy/mean Mean -0.0368429 +trainer/policy/mean Std 0.90783 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76425 +trainer/policy/normal/std Std 0.670026 +trainer/policy/normal/std Max 6.92852 +trainer/policy/normal/std Min 0.29123 +trainer/policy/normal/log_std Mean 0.972227 +trainer/policy/normal/log_std Std 0.343195 +trainer/policy/normal/log_std Max 1.93565 +trainer/policy/normal/log_std Min -1.23364 +eval/num steps total 872320 +eval/num paths total 876 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.065475 +eval/Actions Std 0.9061 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47184 +time/logging (s) 0.00370166 +time/sampling batch (s) 0.28367 +time/saving (s) 0.0033987 +time/training (s) 6.62553 +time/epoch (s) 9.38814 +time/total (s) 8915.58 +Epoch -125 +---------------------------------- --------------- +2022-05-10 15:39:35.302022 PDT | [0] Epoch -124 finished +---------------------------------- --------------- +epoch -124 +replay_buffer/size 999033 +trainer/num train calls 877000 +trainer/Policy Loss -20.5202 +trainer/Log Pis Mean 23.7982 +trainer/Log Pis Std 13.5799 +trainer/Log Pis Max 75.3541 +trainer/Log Pis Min -9.09026 +trainer/policy/mean Mean -0.0303204 +trainer/policy/mean Std 0.908592 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79088 +trainer/policy/normal/std Std 0.666244 +trainer/policy/normal/std Max 6.14043 +trainer/policy/normal/std Min 0.274217 +trainer/policy/normal/log_std Mean 0.985376 +trainer/policy/normal/log_std Std 0.322927 +trainer/policy/normal/log_std Max 1.81489 +trainer/policy/normal/log_std Min -1.29383 +eval/num steps total 873320 +eval/num paths total 877 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0145232 +eval/Actions Std 0.905277 +eval/Actions Max 0.999999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53643 +time/logging (s) 0.00368539 +time/sampling batch (s) 0.284371 +time/saving (s) 0.00344752 +time/training (s) 7.31212 +time/epoch (s) 10.1401 +time/total (s) 8925.72 +Epoch -124 +---------------------------------- --------------- +2022-05-10 15:39:45.959475 PDT | [0] Epoch -123 finished +---------------------------------- --------------- +epoch -123 +replay_buffer/size 999033 +trainer/num train calls 878000 +trainer/Policy Loss -17.9982 +trainer/Log Pis Mean 24.7648 +trainer/Log Pis Std 13.4672 +trainer/Log Pis Max 70.8755 +trainer/Log Pis Min -10.2817 +trainer/policy/mean Mean -0.0431846 +trainer/policy/mean Std 0.901118 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.79962 +trainer/policy/normal/std Std 0.665832 +trainer/policy/normal/std Max 5.46432 +trainer/policy/normal/std Min 0.281617 +trainer/policy/normal/log_std Mean 0.98839 +trainer/policy/normal/log_std Std 0.324657 +trainer/policy/normal/log_std Max 1.69824 +trainer/policy/normal/log_std Min -1.26721 +eval/num steps total 874320 +eval/num paths total 878 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0534799 +eval/Actions Std 0.90157 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64938 +time/logging (s) 0.00421766 +time/sampling batch (s) 0.284599 +time/saving (s) 0.0042851 +time/training (s) 7.69131 +time/epoch (s) 10.6338 +time/total (s) 8936.36 +Epoch -123 +---------------------------------- --------------- +2022-05-10 15:39:55.894028 PDT | [0] Epoch -122 finished +---------------------------------- ---------------- +epoch -122 +replay_buffer/size 999033 +trainer/num train calls 879000 +trainer/Policy Loss -19.4383 +trainer/Log Pis Mean 24.0097 +trainer/Log Pis Std 12.8954 +trainer/Log Pis Max 63.0903 +trainer/Log Pis Min -7.95507 +trainer/policy/mean Mean -0.0339454 +trainer/policy/mean Std 0.90355 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.77421 +trainer/policy/normal/std Std 0.676705 +trainer/policy/normal/std Max 6.712 +trainer/policy/normal/std Min 0.207025 +trainer/policy/normal/log_std Mean 0.976134 +trainer/policy/normal/log_std Std 0.341068 +trainer/policy/normal/log_std Max 1.9039 +trainer/policy/normal/log_std Min -1.57492 +eval/num steps total 875320 +eval/num paths total 879 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 5.30233e-05 +eval/Actions Std 0.925791 +eval/Actions Max 0.999995 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61121 +time/logging (s) 0.00376672 +time/sampling batch (s) 0.528477 +time/saving (s) 0.00362825 +time/training (s) 6.76276 +time/epoch (s) 9.90984 +time/total (s) 8946.27 +Epoch -122 +---------------------------------- ---------------- +2022-05-10 15:40:06.142708 PDT | [0] Epoch -121 finished +---------------------------------- --------------- +epoch -121 +replay_buffer/size 999033 +trainer/num train calls 880000 +trainer/Policy Loss -20.1002 +trainer/Log Pis Mean 24.2011 +trainer/Log Pis Std 13.8496 +trainer/Log Pis Max 73.5953 +trainer/Log Pis Min -10.8243 +trainer/policy/mean Mean -0.0368043 +trainer/policy/mean Std 0.90691 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.71964 +trainer/policy/normal/std Std 0.669287 +trainer/policy/normal/std Max 5.56128 +trainer/policy/normal/std Min 0.226994 +trainer/policy/normal/log_std Mean 0.955392 +trainer/policy/normal/log_std Std 0.34221 +trainer/policy/normal/log_std Max 1.71583 +trainer/policy/normal/log_std Min -1.48283 +eval/num steps total 876320 +eval/num paths total 880 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0156917 +eval/Actions Std 0.892946 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52569 +time/logging (s) 0.00372629 +time/sampling batch (s) 0.536381 +time/saving (s) 0.00355045 +time/training (s) 7.15498 +time/epoch (s) 10.2243 +time/total (s) 8956.5 +Epoch -121 +---------------------------------- --------------- +2022-05-10 15:40:16.827037 PDT | [0] Epoch -120 finished +---------------------------------- --------------- +epoch -120 +replay_buffer/size 999033 +trainer/num train calls 881000 +trainer/Policy Loss -20.2492 +trainer/Log Pis Mean 24.7452 +trainer/Log Pis Std 13.0624 +trainer/Log Pis Max 70.6491 +trainer/Log Pis Min -10.7708 +trainer/policy/mean Mean -0.0268967 +trainer/policy/mean Std 0.909751 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78159 +trainer/policy/normal/std Std 0.671567 +trainer/policy/normal/std Max 5.49909 +trainer/policy/normal/std Min 0.301956 +trainer/policy/normal/log_std Mean 0.97966 +trainer/policy/normal/log_std Std 0.335038 +trainer/policy/normal/log_std Max 1.70458 +trainer/policy/normal/log_std Min -1.19747 +eval/num steps total 877320 +eval/num paths total 881 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.408068 +eval/Actions Std 0.783428 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45265 +time/logging (s) 0.00368904 +time/sampling batch (s) 0.785208 +time/saving (s) 0.00344486 +time/training (s) 7.41528 +time/epoch (s) 10.6603 +time/total (s) 8967.16 +Epoch -120 +---------------------------------- --------------- +2022-05-10 15:40:26.797222 PDT | [0] Epoch -119 finished +---------------------------------- --------------- +epoch -119 +replay_buffer/size 999033 +trainer/num train calls 882000 +trainer/Policy Loss -20.355 +trainer/Log Pis Mean 24.5593 +trainer/Log Pis Std 12.985 +trainer/Log Pis Max 74.6718 +trainer/Log Pis Min -5.83215 +trainer/policy/mean Mean -0.0342313 +trainer/policy/mean Std 0.909432 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.86453 +trainer/policy/normal/std Std 0.677002 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.302206 +trainer/policy/normal/log_std Mean 1.01337 +trainer/policy/normal/log_std Std 0.314007 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19665 +eval/num steps total 878320 +eval/num paths total 882 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0130084 +eval/Actions Std 0.904428 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.30121 +time/logging (s) 0.00384446 +time/sampling batch (s) 0.529472 +time/saving (s) 0.00348969 +time/training (s) 7.10858 +time/epoch (s) 9.9466 +time/total (s) 8977.11 +Epoch -119 +---------------------------------- --------------- +2022-05-10 15:40:36.802287 PDT | [0] Epoch -118 finished +---------------------------------- --------------- +epoch -118 +replay_buffer/size 999033 +trainer/num train calls 883000 +trainer/Policy Loss -21.085 +trainer/Log Pis Mean 24.0565 +trainer/Log Pis Std 12.6748 +trainer/Log Pis Max 67.1621 +trainer/Log Pis Min -9.86745 +trainer/policy/mean Mean -0.0292268 +trainer/policy/mean Std 0.908893 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.785 +trainer/policy/normal/std Std 0.677907 +trainer/policy/normal/std Max 6.68427 +trainer/policy/normal/std Min 0.234687 +trainer/policy/normal/log_std Mean 0.98168 +trainer/policy/normal/log_std Std 0.330557 +trainer/policy/normal/log_std Max 1.89976 +trainer/policy/normal/log_std Min -1.4495 +eval/num steps total 879320 +eval/num paths total 883 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.227557 +eval/Actions Std 0.889187 +eval/Actions Max 0.999994 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54456 +time/logging (s) 0.00366333 +time/sampling batch (s) 0.279876 +time/saving (s) 0.00357127 +time/training (s) 7.14912 +time/epoch (s) 9.98079 +time/total (s) 8987.1 +Epoch -118 +---------------------------------- --------------- +2022-05-10 15:40:47.066632 PDT | [0] Epoch -117 finished +---------------------------------- --------------- +epoch -117 +replay_buffer/size 999033 +trainer/num train calls 884000 +trainer/Policy Loss -19.5301 +trainer/Log Pis Mean 24.9212 +trainer/Log Pis Std 12.8923 +trainer/Log Pis Max 68.3098 +trainer/Log Pis Min -8.22655 +trainer/policy/mean Mean -0.0313706 +trainer/policy/mean Std 0.909964 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.83388 +trainer/policy/normal/std Std 0.666012 +trainer/policy/normal/std Max 5.49753 +trainer/policy/normal/std Min 0.265755 +trainer/policy/normal/log_std Mean 1.00149 +trainer/policy/normal/log_std Std 0.32172 +trainer/policy/normal/log_std Max 1.7043 +trainer/policy/normal/log_std Min -1.32518 +eval/num steps total 880320 +eval/num paths total 884 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00124924 +eval/Actions Std 0.905297 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47062 +time/logging (s) 0.00411124 +time/sampling batch (s) 0.283036 +time/saving (s) 0.00418413 +time/training (s) 7.47887 +time/epoch (s) 10.2408 +time/total (s) 8997.34 +Epoch -117 +---------------------------------- --------------- +2022-05-10 15:40:56.946319 PDT | [0] Epoch -116 finished +---------------------------------- --------------- +epoch -116 +replay_buffer/size 999033 +trainer/num train calls 885000 +trainer/Policy Loss -20.4143 +trainer/Log Pis Mean 23.9022 +trainer/Log Pis Std 13.916 +trainer/Log Pis Max 62.6847 +trainer/Log Pis Min -9.02075 +trainer/policy/mean Mean -0.0388889 +trainer/policy/mean Std 0.906501 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79878 +trainer/policy/normal/std Std 0.65789 +trainer/policy/normal/std Max 6.17619 +trainer/policy/normal/std Min 0.287728 +trainer/policy/normal/log_std Mean 0.988923 +trainer/policy/normal/log_std Std 0.322364 +trainer/policy/normal/log_std Max 1.8207 +trainer/policy/normal/log_std Min -1.24574 +eval/num steps total 881320 +eval/num paths total 885 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.137161 +eval/Actions Std 0.89528 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50584 +time/logging (s) 0.0038263 +time/sampling batch (s) 0.281285 +time/saving (s) 0.00348838 +time/training (s) 7.0607 +time/epoch (s) 9.85514 +time/total (s) 9007.2 +Epoch -116 +---------------------------------- --------------- +2022-05-10 15:41:06.018899 PDT | [0] Epoch -115 finished +---------------------------------- --------------- +epoch -115 +replay_buffer/size 999033 +trainer/num train calls 886000 +trainer/Policy Loss -20.5168 +trainer/Log Pis Mean 25.2083 +trainer/Log Pis Std 13.5463 +trainer/Log Pis Max 74.6225 +trainer/Log Pis Min -5.71293 +trainer/policy/mean Mean -0.0496809 +trainer/policy/mean Std 0.907207 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.83979 +trainer/policy/normal/std Std 0.67781 +trainer/policy/normal/std Max 5.91602 +trainer/policy/normal/std Min 0.24547 +trainer/policy/normal/log_std Mean 1.00153 +trainer/policy/normal/log_std Std 0.331051 +trainer/policy/normal/log_std Max 1.77766 +trainer/policy/normal/log_std Min -1.40458 +eval/num steps total 882320 +eval/num paths total 886 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.149457 +eval/Actions Std 0.89481 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.31713 +time/logging (s) 0.00369103 +time/sampling batch (s) 0.286141 +time/saving (s) 0.00346031 +time/training (s) 6.43783 +time/epoch (s) 9.04825 +time/total (s) 9016.25 +Epoch -115 +---------------------------------- --------------- +2022-05-10 15:41:15.378038 PDT | [0] Epoch -114 finished +---------------------------------- --------------- +epoch -114 +replay_buffer/size 999033 +trainer/num train calls 887000 +trainer/Policy Loss -19.8215 +trainer/Log Pis Mean 24.6386 +trainer/Log Pis Std 12.8055 +trainer/Log Pis Max 65.7087 +trainer/Log Pis Min -3.0559 +trainer/policy/mean Mean -0.0296701 +trainer/policy/mean Std 0.900709 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80424 +trainer/policy/normal/std Std 0.671351 +trainer/policy/normal/std Max 6.14272 +trainer/policy/normal/std Min 0.287428 +trainer/policy/normal/log_std Mean 0.989599 +trainer/policy/normal/log_std Std 0.326311 +trainer/policy/normal/log_std Max 1.81527 +trainer/policy/normal/log_std Min -1.24678 +eval/num steps total 883320 +eval/num paths total 887 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0596724 +eval/Actions Std 0.918198 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48927 +time/logging (s) 0.00415564 +time/sampling batch (s) 0.279728 +time/saving (s) 0.00418496 +time/training (s) 6.55831 +time/epoch (s) 9.33565 +time/total (s) 9025.59 +Epoch -114 +---------------------------------- --------------- +2022-05-10 15:41:24.555235 PDT | [0] Epoch -113 finished +---------------------------------- --------------- +epoch -113 +replay_buffer/size 999033 +trainer/num train calls 888000 +trainer/Policy Loss -20.3938 +trainer/Log Pis Mean 25.0948 +trainer/Log Pis Std 13.4494 +trainer/Log Pis Max 63.5348 +trainer/Log Pis Min -6.28621 +trainer/policy/mean Mean -0.0376843 +trainer/policy/mean Std 0.906022 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.80562 +trainer/policy/normal/std Std 0.679252 +trainer/policy/normal/std Max 5.66142 +trainer/policy/normal/std Min 0.22935 +trainer/policy/normal/log_std Mean 0.986852 +trainer/policy/normal/log_std Std 0.3441 +trainer/policy/normal/log_std Max 1.73368 +trainer/policy/normal/log_std Min -1.47251 +eval/num steps total 884320 +eval/num paths total 888 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0811744 +eval/Actions Std 0.872965 +eval/Actions Max 0.999995 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71446 +time/logging (s) 0.00388184 +time/sampling batch (s) 0.284206 +time/saving (s) 0.00362649 +time/training (s) 6.14587 +time/epoch (s) 9.15205 +time/total (s) 9034.74 +Epoch -113 +---------------------------------- --------------- +2022-05-10 15:41:33.838354 PDT | [0] Epoch -112 finished +---------------------------------- --------------- +epoch -112 +replay_buffer/size 999033 +trainer/num train calls 889000 +trainer/Policy Loss -19.9517 +trainer/Log Pis Mean 24.5492 +trainer/Log Pis Std 14.266 +trainer/Log Pis Max 68.4393 +trainer/Log Pis Min -10.3703 +trainer/policy/mean Mean -0.045827 +trainer/policy/mean Std 0.906473 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81157 +trainer/policy/normal/std Std 0.682158 +trainer/policy/normal/std Max 6.05682 +trainer/policy/normal/std Min 0.220721 +trainer/policy/normal/log_std Mean 0.989902 +trainer/policy/normal/log_std Std 0.338494 +trainer/policy/normal/log_std Max 1.80118 +trainer/policy/normal/log_std Min -1.51086 +eval/num steps total 885320 +eval/num paths total 889 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.117274 +eval/Actions Std 0.903812 +eval/Actions Max 0.999976 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46855 +time/logging (s) 0.00372129 +time/sampling batch (s) 0.281694 +time/saving (s) 0.00341274 +time/training (s) 6.50144 +time/epoch (s) 9.25882 +time/total (s) 9044 +Epoch -112 +---------------------------------- --------------- +2022-05-10 15:41:43.343028 PDT | [0] Epoch -111 finished +---------------------------------- --------------- +epoch -111 +replay_buffer/size 999033 +trainer/num train calls 890000 +trainer/Policy Loss -20.5901 +trainer/Log Pis Mean 24.4764 +trainer/Log Pis Std 14.1808 +trainer/Log Pis Max 72.6975 +trainer/Log Pis Min -6.0982 +trainer/policy/mean Mean -0.0318451 +trainer/policy/mean Std 0.905033 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.82456 +trainer/policy/normal/std Std 0.667482 +trainer/policy/normal/std Max 6.30555 +trainer/policy/normal/std Min 0.293828 +trainer/policy/normal/log_std Mean 0.998177 +trainer/policy/normal/log_std Std 0.320014 +trainer/policy/normal/log_std Max 1.84143 +trainer/policy/normal/log_std Min -1.22476 +eval/num steps total 886320 +eval/num paths total 890 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.175551 +eval/Actions Std 0.905213 +eval/Actions Max 0.999996 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73719 +time/logging (s) 0.00372302 +time/sampling batch (s) 0.282074 +time/saving (s) 0.00354968 +time/training (s) 6.45406 +time/epoch (s) 9.48059 +time/total (s) 9053.49 +Epoch -111 +---------------------------------- --------------- +2022-05-10 15:41:53.294523 PDT | [0] Epoch -110 finished +---------------------------------- --------------- +epoch -110 +replay_buffer/size 999033 +trainer/num train calls 891000 +trainer/Policy Loss -20.0198 +trainer/Log Pis Mean 24.0829 +trainer/Log Pis Std 12.5561 +trainer/Log Pis Max 69.7529 +trainer/Log Pis Min -5.84961 +trainer/policy/mean Mean -0.0300411 +trainer/policy/mean Std 0.903172 +trainer/policy/mean Max 0.99997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.83159 +trainer/policy/normal/std Std 0.685259 +trainer/policy/normal/std Max 5.97321 +trainer/policy/normal/std Min 0.286776 +trainer/policy/normal/log_std Mean 0.997824 +trainer/policy/normal/log_std Std 0.3334 +trainer/policy/normal/log_std Max 1.78728 +trainer/policy/normal/log_std Min -1.24905 +eval/num steps total 887320 +eval/num paths total 891 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.00933961 +eval/Actions Std 0.909465 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52172 +time/logging (s) 0.00370793 +time/sampling batch (s) 0.278713 +time/saving (s) 0.00345282 +time/training (s) 7.12007 +time/epoch (s) 9.92766 +time/total (s) 9063.42 +Epoch -110 +---------------------------------- --------------- +2022-05-10 15:42:03.046187 PDT | [0] Epoch -109 finished +---------------------------------- --------------- +epoch -109 +replay_buffer/size 999033 +trainer/num train calls 892000 +trainer/Policy Loss -19.9435 +trainer/Log Pis Mean 23.4286 +trainer/Log Pis Std 13.9786 +trainer/Log Pis Max 81.5259 +trainer/Log Pis Min -9.80689 +trainer/policy/mean Mean -0.0139109 +trainer/policy/mean Std 0.907561 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78417 +trainer/policy/normal/std Std 0.686207 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.243966 +trainer/policy/normal/log_std Mean 0.980073 +trainer/policy/normal/log_std Std 0.335102 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.41073 +eval/num steps total 888320 +eval/num paths total 892 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.10148 +eval/Actions Std 0.901765 +eval/Actions Max 0.999982 +eval/Actions Min -0.999982 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57171 +time/logging (s) 0.00425317 +time/sampling batch (s) 0.281462 +time/saving (s) 0.00422185 +time/training (s) 6.86628 +time/epoch (s) 9.72793 +time/total (s) 9073.15 +Epoch -109 +---------------------------------- --------------- +2022-05-10 15:42:12.638220 PDT | [0] Epoch -108 finished +---------------------------------- --------------- +epoch -108 +replay_buffer/size 999033 +trainer/num train calls 893000 +trainer/Policy Loss -19.6617 +trainer/Log Pis Mean 23.8303 +trainer/Log Pis Std 13.3374 +trainer/Log Pis Max 74.98 +trainer/Log Pis Min -9.79478 +trainer/policy/mean Mean -0.0252164 +trainer/policy/mean Std 0.903927 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.80951 +trainer/policy/normal/std Std 0.668135 +trainer/policy/normal/std Max 6.61963 +trainer/policy/normal/std Min 0.278293 +trainer/policy/normal/log_std Mean 0.991637 +trainer/policy/normal/log_std Std 0.326622 +trainer/policy/normal/log_std Max 1.89004 +trainer/policy/normal/log_std Min -1.27908 +eval/num steps total 889320 +eval/num paths total 893 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.2014 +eval/Actions Std 0.910032 +eval/Actions Max 0.99999 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69791 +time/logging (s) 0.00370138 +time/sampling batch (s) 0.531155 +time/saving (s) 0.00370775 +time/training (s) 6.33029 +time/epoch (s) 9.56676 +time/total (s) 9082.72 +Epoch -108 +---------------------------------- --------------- +2022-05-10 15:42:22.378722 PDT | [0] Epoch -107 finished +---------------------------------- --------------- +epoch -107 +replay_buffer/size 999033 +trainer/num train calls 894000 +trainer/Policy Loss -18.8021 +trainer/Log Pis Mean 24.1936 +trainer/Log Pis Std 12.9251 +trainer/Log Pis Max 71.2624 +trainer/Log Pis Min -6.40543 +trainer/policy/mean Mean -0.0316946 +trainer/policy/mean Std 0.907214 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.81308 +trainer/policy/normal/std Std 0.676462 +trainer/policy/normal/std Max 5.98849 +trainer/policy/normal/std Min 0.293861 +trainer/policy/normal/log_std Mean 0.99295 +trainer/policy/normal/log_std Std 0.324949 +trainer/policy/normal/log_std Max 1.78984 +trainer/policy/normal/log_std Min -1.22465 +eval/num steps total 890320 +eval/num paths total 894 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0889306 +eval/Actions Std 0.904442 +eval/Actions Max 0.999995 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49129 +time/logging (s) 0.00368459 +time/sampling batch (s) 0.281284 +time/saving (s) 0.00344967 +time/training (s) 6.93673 +time/epoch (s) 9.71644 +time/total (s) 9092.44 +Epoch -107 +---------------------------------- --------------- +2022-05-10 15:42:31.986506 PDT | [0] Epoch -106 finished +---------------------------------- --------------- +epoch -106 +replay_buffer/size 999033 +trainer/num train calls 895000 +trainer/Policy Loss -19.0562 +trainer/Log Pis Mean 24.507 +trainer/Log Pis Std 13.584 +trainer/Log Pis Max 84.8925 +trainer/Log Pis Min -7.27013 +trainer/policy/mean Mean -0.0359926 +trainer/policy/mean Std 0.902275 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.7593 +trainer/policy/normal/std Std 0.66237 +trainer/policy/normal/std Max 5.54444 +trainer/policy/normal/std Min 0.231851 +trainer/policy/normal/log_std Mean 0.973334 +trainer/policy/normal/log_std Std 0.326323 +trainer/policy/normal/log_std Max 1.7128 +trainer/policy/normal/log_std Min -1.46166 +eval/num steps total 891320 +eval/num paths total 895 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.106682 +eval/Actions Std 0.901813 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52964 +time/logging (s) 0.00363369 +time/sampling batch (s) 0.280641 +time/saving (s) 0.00349708 +time/training (s) 6.76644 +time/epoch (s) 9.58385 +time/total (s) 9102.03 +Epoch -106 +---------------------------------- --------------- +2022-05-10 15:42:41.740815 PDT | [0] Epoch -105 finished +---------------------------------- --------------- +epoch -105 +replay_buffer/size 999033 +trainer/num train calls 896000 +trainer/Policy Loss -18.586 +trainer/Log Pis Mean 23.3896 +trainer/Log Pis Std 12.9462 +trainer/Log Pis Max 69.3405 +trainer/Log Pis Min -6.35078 +trainer/policy/mean Mean -0.0154741 +trainer/policy/mean Std 0.906285 +trainer/policy/mean Max 0.999975 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.78545 +trainer/policy/normal/std Std 0.66473 +trainer/policy/normal/std Max 6.17098 +trainer/policy/normal/std Min 0.203608 +trainer/policy/normal/log_std Mean 0.981676 +trainer/policy/normal/log_std Std 0.334578 +trainer/policy/normal/log_std Max 1.81986 +trainer/policy/normal/log_std Min -1.59156 +eval/num steps total 892320 +eval/num paths total 896 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0858198 +eval/Actions Std 0.894047 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51369 +time/logging (s) 0.00383302 +time/sampling batch (s) 0.534104 +time/saving (s) 0.00351569 +time/training (s) 6.67536 +time/epoch (s) 9.7305 +time/total (s) 9111.76 +Epoch -105 +---------------------------------- --------------- +2022-05-10 15:42:51.219605 PDT | [0] Epoch -104 finished +---------------------------------- --------------- +epoch -104 +replay_buffer/size 999033 +trainer/num train calls 897000 +trainer/Policy Loss -19.6714 +trainer/Log Pis Mean 23.1928 +trainer/Log Pis Std 12.7952 +trainer/Log Pis Max 69.2692 +trainer/Log Pis Min -11.945 +trainer/policy/mean Mean -0.0469499 +trainer/policy/mean Std 0.908806 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.75786 +trainer/policy/normal/std Std 0.645638 +trainer/policy/normal/std Max 5.01982 +trainer/policy/normal/std Min 0.271162 +trainer/policy/normal/log_std Mean 0.973433 +trainer/policy/normal/log_std Std 0.325952 +trainer/policy/normal/log_std Max 1.61339 +trainer/policy/normal/log_std Min -1.30504 +eval/num steps total 893320 +eval/num paths total 897 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114318 +eval/Actions Std 0.913952 +eval/Actions Max 0.999996 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.75407 +time/logging (s) 0.0037931 +time/sampling batch (s) 0.28091 +time/saving (s) 0.00345227 +time/training (s) 6.41231 +time/epoch (s) 9.45453 +time/total (s) 9121.22 +Epoch -104 +---------------------------------- --------------- +2022-05-10 15:43:01.371318 PDT | [0] Epoch -103 finished +---------------------------------- --------------- +epoch -103 +replay_buffer/size 999033 +trainer/num train calls 898000 +trainer/Policy Loss -21.0728 +trainer/Log Pis Mean 23.9981 +trainer/Log Pis Std 13.2904 +trainer/Log Pis Max 60.5883 +trainer/Log Pis Min -9.90423 +trainer/policy/mean Mean -0.0325717 +trainer/policy/mean Std 0.906093 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79502 +trainer/policy/normal/std Std 0.69531 +trainer/policy/normal/std Max 5.86607 +trainer/policy/normal/std Min 0.261381 +trainer/policy/normal/log_std Mean 0.980442 +trainer/policy/normal/log_std Std 0.354328 +trainer/policy/normal/log_std Max 1.76918 +trainer/policy/normal/log_std Min -1.34177 +eval/num steps total 894320 +eval/num paths total 898 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.320532 +eval/Actions Std 0.868983 +eval/Actions Max 0.999997 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67025 +time/logging (s) 0.00395706 +time/sampling batch (s) 0.283121 +time/saving (s) 0.00403651 +time/training (s) 7.16628 +time/epoch (s) 10.1276 +time/total (s) 9131.35 +Epoch -103 +---------------------------------- --------------- +2022-05-10 15:43:12.079243 PDT | [0] Epoch -102 finished +---------------------------------- --------------- +epoch -102 +replay_buffer/size 999033 +trainer/num train calls 899000 +trainer/Policy Loss -20.0898 +trainer/Log Pis Mean 24.0859 +trainer/Log Pis Std 13.02 +trainer/Log Pis Max 65.8514 +trainer/Log Pis Min -2.13843 +trainer/policy/mean Mean -0.0254127 +trainer/policy/mean Std 0.911888 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.73607 +trainer/policy/normal/std Std 0.658446 +trainer/policy/normal/std Max 5.71324 +trainer/policy/normal/std Min 0.216319 +trainer/policy/normal/log_std Mean 0.962786 +trainer/policy/normal/log_std Std 0.338282 +trainer/policy/normal/log_std Max 1.74279 +trainer/policy/normal/log_std Min -1.531 +eval/num steps total 895320 +eval/num paths total 899 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123379 +eval/Actions Std 0.943319 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57408 +time/logging (s) 0.00383636 +time/sampling batch (s) 0.531276 +time/saving (s) 0.00361554 +time/training (s) 7.57072 +time/epoch (s) 10.6835 +time/total (s) 9142.03 +Epoch -102 +---------------------------------- --------------- +2022-05-10 15:43:22.167570 PDT | [0] Epoch -101 finished +---------------------------------- --------------- +epoch -101 +replay_buffer/size 999033 +trainer/num train calls 900000 +trainer/Policy Loss -19.2567 +trainer/Log Pis Mean 24.9158 +trainer/Log Pis Std 12.4616 +trainer/Log Pis Max 70.9872 +trainer/Log Pis Min -3.98007 +trainer/policy/mean Mean -0.0295063 +trainer/policy/mean Std 0.905527 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.79506 +trainer/policy/normal/std Std 0.695549 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.267712 +trainer/policy/normal/log_std Mean 0.98126 +trainer/policy/normal/log_std Std 0.349795 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.31784 +eval/num steps total 896320 +eval/num paths total 900 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135982 +eval/Actions Std 0.89266 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73138 +time/logging (s) 0.00378571 +time/sampling batch (s) 0.283201 +time/saving (s) 0.00652083 +time/training (s) 7.03934 +time/epoch (s) 10.0642 +time/total (s) 9152.1 +Epoch -101 +---------------------------------- --------------- +2022-05-10 15:43:32.551591 PDT | [0] Epoch -100 finished +---------------------------------- --------------- +epoch -100 +replay_buffer/size 999033 +trainer/num train calls 901000 +trainer/Policy Loss -19.6574 +trainer/Log Pis Mean 24.9429 +trainer/Log Pis Std 13.4221 +trainer/Log Pis Max 63.7315 +trainer/Log Pis Min -6.81643 +trainer/policy/mean Mean -0.0593477 +trainer/policy/mean Std 0.906286 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79617 +trainer/policy/normal/std Std 0.675984 +trainer/policy/normal/std Max 6.55486 +trainer/policy/normal/std Min 0.242416 +trainer/policy/normal/log_std Mean 0.984114 +trainer/policy/normal/log_std Std 0.340601 +trainer/policy/normal/log_std Max 1.88021 +trainer/policy/normal/log_std Min -1.4171 +eval/num steps total 897320 +eval/num paths total 901 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.159152 +eval/Actions Std 0.920551 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72855 +time/logging (s) 0.00381734 +time/sampling batch (s) 0.530023 +time/saving (s) 0.00345856 +time/training (s) 7.09439 +time/epoch (s) 10.3602 +time/total (s) 9162.46 +Epoch -100 +---------------------------------- --------------- +2022-05-10 15:43:42.922367 PDT | [0] Epoch -99 finished +---------------------------------- --------------- +epoch -99 +replay_buffer/size 999033 +trainer/num train calls 902000 +trainer/Policy Loss -20.4011 +trainer/Log Pis Mean 25.1377 +trainer/Log Pis Std 13.9003 +trainer/Log Pis Max 74.3181 +trainer/Log Pis Min -6.6097 +trainer/policy/mean Mean -0.0377608 +trainer/policy/mean Std 0.906914 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.70262 +trainer/policy/normal/std Std 0.673311 +trainer/policy/normal/std Max 5.77325 +trainer/policy/normal/std Min 0.263845 +trainer/policy/normal/log_std Mean 0.946942 +trainer/policy/normal/log_std Std 0.3519 +trainer/policy/normal/log_std Max 1.75324 +trainer/policy/normal/log_std Min -1.33239 +eval/num steps total 898320 +eval/num paths total 902 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.131824 +eval/Actions Std 0.883606 +eval/Actions Max 0.99999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77818 +time/logging (s) 0.00386013 +time/sampling batch (s) 0.531373 +time/saving (s) 0.00347959 +time/training (s) 7.03007 +time/epoch (s) 10.347 +time/total (s) 9172.81 +Epoch -99 +---------------------------------- --------------- +2022-05-10 15:43:52.510180 PDT | [0] Epoch -98 finished +---------------------------------- --------------- +epoch -98 +replay_buffer/size 999033 +trainer/num train calls 903000 +trainer/Policy Loss -20.4521 +trainer/Log Pis Mean 24.665 +trainer/Log Pis Std 13.3038 +trainer/Log Pis Max 70.8007 +trainer/Log Pis Min -6.78768 +trainer/policy/mean Mean -0.03449 +trainer/policy/mean Std 0.907325 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.83364 +trainer/policy/normal/std Std 0.671848 +trainer/policy/normal/std Max 6.53245 +trainer/policy/normal/std Min 0.237873 +trainer/policy/normal/log_std Mean 1.00083 +trainer/policy/normal/log_std Std 0.323549 +trainer/policy/normal/log_std Max 1.87678 +trainer/policy/normal/log_std Min -1.43602 +eval/num steps total 899320 +eval/num paths total 903 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112836 +eval/Actions Std 0.851071 +eval/Actions Max 0.999999 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84637 +time/logging (s) 0.00385278 +time/sampling batch (s) 0.278907 +time/saving (s) 0.00344173 +time/training (s) 6.43145 +time/epoch (s) 9.56402 +time/total (s) 9182.38 +Epoch -98 +---------------------------------- --------------- +2022-05-10 15:44:04.082883 PDT | [0] Epoch -97 finished +---------------------------------- --------------- +epoch -97 +replay_buffer/size 999033 +trainer/num train calls 904000 +trainer/Policy Loss -19.0678 +trainer/Log Pis Mean 24.377 +trainer/Log Pis Std 12.9297 +trainer/Log Pis Max 61.738 +trainer/Log Pis Min -8.26092 +trainer/policy/mean Mean -0.0446977 +trainer/policy/mean Std 0.908945 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77323 +trainer/policy/normal/std Std 0.691068 +trainer/policy/normal/std Max 5.41498 +trainer/policy/normal/std Min 0.25618 +trainer/policy/normal/log_std Mean 0.971641 +trainer/policy/normal/log_std Std 0.359682 +trainer/policy/normal/log_std Max 1.68917 +trainer/policy/normal/log_std Min -1.36187 +eval/num steps total 900320 +eval/num paths total 904 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.305605 +eval/Actions Std 0.878933 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76578 +time/logging (s) 0.00384728 +time/sampling batch (s) 0.795524 +time/saving (s) 0.00344768 +time/training (s) 7.97928 +time/epoch (s) 11.5479 +time/total (s) 9193.93 +Epoch -97 +---------------------------------- --------------- +2022-05-10 15:44:13.824775 PDT | [0] Epoch -96 finished +---------------------------------- --------------- +epoch -96 +replay_buffer/size 999033 +trainer/num train calls 905000 +trainer/Policy Loss -19.3617 +trainer/Log Pis Mean 24.6265 +trainer/Log Pis Std 13.3983 +trainer/Log Pis Max 71.4221 +trainer/Log Pis Min -6.2508 +trainer/policy/mean Mean -0.0527051 +trainer/policy/mean Std 0.907759 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.77504 +trainer/policy/normal/std Std 0.652017 +trainer/policy/normal/std Max 5.89667 +trainer/policy/normal/std Min 0.353718 +trainer/policy/normal/log_std Mean 0.982454 +trainer/policy/normal/log_std Std 0.308644 +trainer/policy/normal/log_std Max 1.77439 +trainer/policy/normal/log_std Min -1.03926 +eval/num steps total 901320 +eval/num paths total 905 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0512605 +eval/Actions Std 0.873924 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52679 +time/logging (s) 0.00376869 +time/sampling batch (s) 0.28777 +time/saving (s) 0.0034916 +time/training (s) 6.89558 +time/epoch (s) 9.7174 +time/total (s) 9203.65 +Epoch -96 +---------------------------------- --------------- +2022-05-10 15:44:23.136632 PDT | [0] Epoch -95 finished +---------------------------------- --------------- +epoch -95 +replay_buffer/size 999033 +trainer/num train calls 906000 +trainer/Policy Loss -19.0726 +trainer/Log Pis Mean 24.2958 +trainer/Log Pis Std 13.1205 +trainer/Log Pis Max 61.8281 +trainer/Log Pis Min -8.3637 +trainer/policy/mean Mean -0.0438816 +trainer/policy/mean Std 0.903406 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85983 +trainer/policy/normal/std Std 0.701437 +trainer/policy/normal/std Max 6.32351 +trainer/policy/normal/std Min 0.310249 +trainer/policy/normal/log_std Mean 1.00505 +trainer/policy/normal/log_std Std 0.346946 +trainer/policy/normal/log_std Max 1.84427 +trainer/policy/normal/log_std Min -1.17038 +eval/num steps total 902320 +eval/num paths total 906 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.156888 +eval/Actions Std 0.919009 +eval/Actions Max 0.999987 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43094 +time/logging (s) 0.00372516 +time/sampling batch (s) 0.28446 +time/saving (s) 0.00343586 +time/training (s) 6.56458 +time/epoch (s) 9.28714 +time/total (s) 9212.94 +Epoch -95 +---------------------------------- --------------- +2022-05-10 15:44:33.024283 PDT | [0] Epoch -94 finished +---------------------------------- --------------- +epoch -94 +replay_buffer/size 999033 +trainer/num train calls 907000 +trainer/Policy Loss -19.7939 +trainer/Log Pis Mean 24.6481 +trainer/Log Pis Std 13.1791 +trainer/Log Pis Max 68.5993 +trainer/Log Pis Min -8.88756 +trainer/policy/mean Mean -0.0476438 +trainer/policy/mean Std 0.905324 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.79629 +trainer/policy/normal/std Std 0.683535 +trainer/policy/normal/std Max 6.55273 +trainer/policy/normal/std Min 0.256408 +trainer/policy/normal/log_std Mean 0.983884 +trainer/policy/normal/log_std Std 0.33942 +trainer/policy/normal/log_std Max 1.87988 +trainer/policy/normal/log_std Min -1.36099 +eval/num steps total 903320 +eval/num paths total 907 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0626371 +eval/Actions Std 0.914468 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58833 +time/logging (s) 0.00371953 +time/sampling batch (s) 0.287758 +time/saving (s) 0.00353295 +time/training (s) 6.97997 +time/epoch (s) 9.86332 +time/total (s) 9222.81 +Epoch -94 +---------------------------------- --------------- +2022-05-10 15:44:42.691233 PDT | [0] Epoch -93 finished +---------------------------------- --------------- +epoch -93 +replay_buffer/size 999033 +trainer/num train calls 908000 +trainer/Policy Loss -19.2793 +trainer/Log Pis Mean 24.193 +trainer/Log Pis Std 13.3222 +trainer/Log Pis Max 67.8386 +trainer/Log Pis Min -13.1294 +trainer/policy/mean Mean -0.0423058 +trainer/policy/mean Std 0.902809 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.71105 +trainer/policy/normal/std Std 0.704742 +trainer/policy/normal/std Max 6.04357 +trainer/policy/normal/std Min 0.284202 +trainer/policy/normal/log_std Mean 0.946108 +trainer/policy/normal/log_std Std 0.367023 +trainer/policy/normal/log_std Max 1.79899 +trainer/policy/normal/log_std Min -1.25807 +eval/num steps total 904320 +eval/num paths total 908 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.22568 +eval/Actions Std 0.932846 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48721 +time/logging (s) 0.00396356 +time/sampling batch (s) 0.281826 +time/saving (s) 0.00391617 +time/training (s) 6.8659 +time/epoch (s) 9.64281 +time/total (s) 9232.45 +Epoch -93 +---------------------------------- --------------- +2022-05-10 15:44:52.353308 PDT | [0] Epoch -92 finished +---------------------------------- --------------- +epoch -92 +replay_buffer/size 999033 +trainer/num train calls 909000 +trainer/Policy Loss -18.8464 +trainer/Log Pis Mean 24.9477 +trainer/Log Pis Std 13.607 +trainer/Log Pis Max 73.3112 +trainer/Log Pis Min -9.01602 +trainer/policy/mean Mean -0.0274535 +trainer/policy/mean Std 0.905367 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.83787 +trainer/policy/normal/std Std 0.684986 +trainer/policy/normal/std Max 5.99364 +trainer/policy/normal/std Min 0.248005 +trainer/policy/normal/log_std Mean 1.0006 +trainer/policy/normal/log_std Std 0.330437 +trainer/policy/normal/log_std Max 1.7907 +trainer/policy/normal/log_std Min -1.39431 +eval/num steps total 905320 +eval/num paths total 909 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00391964 +eval/Actions Std 0.895431 +eval/Actions Max 0.999999 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.48452 +time/logging (s) 0.00381118 +time/sampling batch (s) 0.280352 +time/saving (s) 0.00368258 +time/training (s) 6.86534 +time/epoch (s) 9.63771 +time/total (s) 9242.09 +Epoch -92 +---------------------------------- --------------- +2022-05-10 15:45:02.885004 PDT | [0] Epoch -91 finished +---------------------------------- --------------- +epoch -91 +replay_buffer/size 999033 +trainer/num train calls 910000 +trainer/Policy Loss -19.5499 +trainer/Log Pis Mean 25.1314 +trainer/Log Pis Std 13.0767 +trainer/Log Pis Max 72.4342 +trainer/Log Pis Min -10.8748 +trainer/policy/mean Mean -0.0456712 +trainer/policy/mean Std 0.908545 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80846 +trainer/policy/normal/std Std 0.688668 +trainer/policy/normal/std Max 6.4367 +trainer/policy/normal/std Min 0.290254 +trainer/policy/normal/log_std Mean 0.989681 +trainer/policy/normal/log_std Std 0.331712 +trainer/policy/normal/log_std Max 1.86202 +trainer/policy/normal/log_std Min -1.237 +eval/num steps total 906320 +eval/num paths total 910 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.112755 +eval/Actions Std 0.90569 +eval/Actions Max 0.999997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.71647 +time/logging (s) 0.00375681 +time/sampling batch (s) 0.281377 +time/saving (s) 0.00356129 +time/training (s) 7.5021 +time/epoch (s) 10.5073 +time/total (s) 9252.6 +Epoch -91 +---------------------------------- --------------- +2022-05-10 15:45:13.080507 PDT | [0] Epoch -90 finished +---------------------------------- --------------- +epoch -90 +replay_buffer/size 999033 +trainer/num train calls 911000 +trainer/Policy Loss -19.4116 +trainer/Log Pis Mean 24.3789 +trainer/Log Pis Std 13.6116 +trainer/Log Pis Max 77.8378 +trainer/Log Pis Min -10.8411 +trainer/policy/mean Mean -0.0266964 +trainer/policy/mean Std 0.900809 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.78838 +trainer/policy/normal/std Std 0.70533 +trainer/policy/normal/std Max 6.27828 +trainer/policy/normal/std Min 0.271928 +trainer/policy/normal/log_std Mean 0.978877 +trainer/policy/normal/log_std Std 0.345798 +trainer/policy/normal/log_std Max 1.8371 +trainer/policy/normal/log_std Min -1.30222 +eval/num steps total 907320 +eval/num paths total 911 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0565462 +eval/Actions Std 0.881468 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66228 +time/logging (s) 0.00365019 +time/sampling batch (s) 0.280752 +time/saving (s) 0.00344147 +time/training (s) 7.22121 +time/epoch (s) 10.1713 +time/total (s) 9262.78 +Epoch -90 +---------------------------------- --------------- +2022-05-10 15:45:23.984652 PDT | [0] Epoch -89 finished +---------------------------------- --------------- +epoch -89 +replay_buffer/size 999033 +trainer/num train calls 912000 +trainer/Policy Loss -19.6334 +trainer/Log Pis Mean 24.6744 +trainer/Log Pis Std 13.2239 +trainer/Log Pis Max 69.3388 +trainer/Log Pis Min -11.3962 +trainer/policy/mean Mean -0.040283 +trainer/policy/mean Std 0.902006 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.78472 +trainer/policy/normal/std Std 0.688573 +trainer/policy/normal/std Max 6.31222 +trainer/policy/normal/std Min 0.274991 +trainer/policy/normal/log_std Mean 0.979578 +trainer/policy/normal/log_std Std 0.339501 +trainer/policy/normal/log_std Max 1.84249 +trainer/policy/normal/log_std Min -1.29102 +eval/num steps total 908320 +eval/num paths total 912 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0749269 +eval/Actions Std 0.910901 +eval/Actions Max 0.999995 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68848 +time/logging (s) 0.00375052 +time/sampling batch (s) 0.295847 +time/saving (s) 0.00347282 +time/training (s) 7.8879 +time/epoch (s) 10.8794 +time/total (s) 9273.66 +Epoch -89 +---------------------------------- --------------- +2022-05-10 15:45:33.643234 PDT | [0] Epoch -88 finished +---------------------------------- --------------- +epoch -88 +replay_buffer/size 999033 +trainer/num train calls 913000 +trainer/Policy Loss -18.9762 +trainer/Log Pis Mean 24.5943 +trainer/Log Pis Std 13.4637 +trainer/Log Pis Max 69.769 +trainer/Log Pis Min -12.0507 +trainer/policy/mean Mean -0.0348611 +trainer/policy/mean Std 0.908585 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81679 +trainer/policy/normal/std Std 0.641458 +trainer/policy/normal/std Max 7.10453 +trainer/policy/normal/std Min 0.245441 +trainer/policy/normal/log_std Mean 0.999476 +trainer/policy/normal/log_std Std 0.30081 +trainer/policy/normal/log_std Max 1.96073 +trainer/policy/normal/log_std Min -1.4047 +eval/num steps total 909320 +eval/num paths total 913 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.114245 +eval/Actions Std 0.886844 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50524 +time/logging (s) 0.00369517 +time/sampling batch (s) 0.534587 +time/saving (s) 0.00348415 +time/training (s) 6.58709 +time/epoch (s) 9.63409 +time/total (s) 9283.3 +Epoch -88 +---------------------------------- --------------- +2022-05-10 15:45:43.044348 PDT | [0] Epoch -87 finished +---------------------------------- --------------- +epoch -87 +replay_buffer/size 999033 +trainer/num train calls 914000 +trainer/Policy Loss -19.5618 +trainer/Log Pis Mean 24.6764 +trainer/Log Pis Std 13.2667 +trainer/Log Pis Max 65.8247 +trainer/Log Pis Min -7.50966 +trainer/policy/mean Mean -0.0460971 +trainer/policy/mean Std 0.904783 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.73904 +trainer/policy/normal/std Std 0.678567 +trainer/policy/normal/std Max 5.61949 +trainer/policy/normal/std Min 0.2521 +trainer/policy/normal/log_std Mean 0.962532 +trainer/policy/normal/log_std Std 0.340949 +trainer/policy/normal/log_std Max 1.72624 +trainer/policy/normal/log_std Min -1.37793 +eval/num steps total 910320 +eval/num paths total 914 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.261257 +eval/Actions Std 0.758414 +eval/Actions Max 0.999983 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76834 +time/logging (s) 0.00371055 +time/sampling batch (s) 0.284969 +time/saving (s) 0.00359321 +time/training (s) 6.3161 +time/epoch (s) 9.37672 +time/total (s) 9292.68 +Epoch -87 +---------------------------------- --------------- +2022-05-10 15:45:52.509460 PDT | [0] Epoch -86 finished +---------------------------------- --------------- +epoch -86 +replay_buffer/size 999033 +trainer/num train calls 915000 +trainer/Policy Loss -20.5985 +trainer/Log Pis Mean 24.2747 +trainer/Log Pis Std 13.5747 +trainer/Log Pis Max 70.3035 +trainer/Log Pis Min -9.86318 +trainer/policy/mean Mean -0.0348211 +trainer/policy/mean Std 0.906653 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81648 +trainer/policy/normal/std Std 0.6635 +trainer/policy/normal/std Max 6.16311 +trainer/policy/normal/std Min 0.261491 +trainer/policy/normal/log_std Mean 0.995604 +trainer/policy/normal/log_std Std 0.319239 +trainer/policy/normal/log_std Max 1.81858 +trainer/policy/normal/log_std Min -1.34135 +eval/num steps total 911320 +eval/num paths total 915 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.213086 +eval/Actions Std 0.896064 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76378 +time/logging (s) 0.00377048 +time/sampling batch (s) 0.285527 +time/saving (s) 0.00348151 +time/training (s) 6.38416 +time/epoch (s) 9.44072 +time/total (s) 9302.12 +Epoch -86 +---------------------------------- --------------- +2022-05-10 15:46:02.946932 PDT | [0] Epoch -85 finished +---------------------------------- --------------- +epoch -85 +replay_buffer/size 999033 +trainer/num train calls 916000 +trainer/Policy Loss -20.5902 +trainer/Log Pis Mean 24.9661 +trainer/Log Pis Std 13.0943 +trainer/Log Pis Max 65.1822 +trainer/Log Pis Min -2.83033 +trainer/policy/mean Mean -0.043443 +trainer/policy/mean Std 0.905466 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76996 +trainer/policy/normal/std Std 0.692181 +trainer/policy/normal/std Max 5.43311 +trainer/policy/normal/std Min 0.233533 +trainer/policy/normal/log_std Mean 0.968592 +trainer/policy/normal/log_std Std 0.36923 +trainer/policy/normal/log_std Max 1.69251 +trainer/policy/normal/log_std Min -1.45443 +eval/num steps total 912320 +eval/num paths total 916 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.00757616 +eval/Actions Std 0.963322 +eval/Actions Max 0.999997 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.7467 +time/logging (s) 0.00381677 +time/sampling batch (s) 0.284058 +time/saving (s) 0.00345021 +time/training (s) 7.37514 +time/epoch (s) 10.4132 +time/total (s) 9312.54 +Epoch -85 +---------------------------------- --------------- +2022-05-10 15:46:12.714249 PDT | [0] Epoch -84 finished +---------------------------------- --------------- +epoch -84 +replay_buffer/size 999033 +trainer/num train calls 917000 +trainer/Policy Loss -19.767 +trainer/Log Pis Mean 23.8837 +trainer/Log Pis Std 13.4365 +trainer/Log Pis Max 63.0723 +trainer/Log Pis Min -8.25796 +trainer/policy/mean Mean -0.02827 +trainer/policy/mean Std 0.907491 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.8044 +trainer/policy/normal/std Std 0.69403 +trainer/policy/normal/std Max 5.98611 +trainer/policy/normal/std Min 0.240388 +trainer/policy/normal/log_std Mean 0.984265 +trainer/policy/normal/log_std Std 0.353165 +trainer/policy/normal/log_std Max 1.78944 +trainer/policy/normal/log_std Min -1.4255 +eval/num steps total 913320 +eval/num paths total 917 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0604448 +eval/Actions Std 0.842179 +eval/Actions Max 0.999992 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63853 +time/logging (s) 0.00398594 +time/sampling batch (s) 0.286058 +time/saving (s) 0.00399503 +time/training (s) 6.81046 +time/epoch (s) 9.74303 +time/total (s) 9322.28 +Epoch -84 +---------------------------------- --------------- +2022-05-10 15:46:22.282035 PDT | [0] Epoch -83 finished +---------------------------------- --------------- +epoch -83 +replay_buffer/size 999033 +trainer/num train calls 918000 +trainer/Policy Loss -19.1 +trainer/Log Pis Mean 23.9077 +trainer/Log Pis Std 13.6195 +trainer/Log Pis Max 70.0628 +trainer/Log Pis Min -8.96938 +trainer/policy/mean Mean -0.0345299 +trainer/policy/mean Std 0.903636 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.84077 +trainer/policy/normal/std Std 0.704055 +trainer/policy/normal/std Max 6.13129 +trainer/policy/normal/std Min 0.238733 +trainer/policy/normal/log_std Mean 0.99893 +trainer/policy/normal/log_std Std 0.342373 +trainer/policy/normal/log_std Max 1.8134 +trainer/policy/normal/log_std Min -1.43241 +eval/num steps total 914320 +eval/num paths total 918 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.226318 +eval/Actions Std 0.737872 +eval/Actions Max 0.999981 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55301 +time/logging (s) 0.00380769 +time/sampling batch (s) 0.280373 +time/saving (s) 0.00373728 +time/training (s) 6.70245 +time/epoch (s) 9.54338 +time/total (s) 9331.83 +Epoch -83 +---------------------------------- --------------- +2022-05-10 15:46:32.264077 PDT | [0] Epoch -82 finished +---------------------------------- --------------- +epoch -82 +replay_buffer/size 999033 +trainer/num train calls 919000 +trainer/Policy Loss -20.9366 +trainer/Log Pis Mean 24.7211 +trainer/Log Pis Std 13.3038 +trainer/Log Pis Max 62.9557 +trainer/Log Pis Min -9.44505 +trainer/policy/mean Mean -0.0350945 +trainer/policy/mean Std 0.908734 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.74839 +trainer/policy/normal/std Std 0.691039 +trainer/policy/normal/std Max 5.54672 +trainer/policy/normal/std Min 0.273052 +trainer/policy/normal/log_std Mean 0.961848 +trainer/policy/normal/log_std Std 0.362465 +trainer/policy/normal/log_std Max 1.71321 +trainer/policy/normal/log_std Min -1.29809 +eval/num steps total 915320 +eval/num paths total 919 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0946434 +eval/Actions Std 0.91526 +eval/Actions Max 0.999988 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6384 +time/logging (s) 0.00375243 +time/sampling batch (s) 0.282418 +time/saving (s) 0.00351113 +time/training (s) 7.0297 +time/epoch (s) 9.95778 +time/total (s) 9341.79 +Epoch -82 +---------------------------------- --------------- +2022-05-10 15:46:42.673075 PDT | [0] Epoch -81 finished +---------------------------------- --------------- +epoch -81 +replay_buffer/size 999033 +trainer/num train calls 920000 +trainer/Policy Loss -19.6678 +trainer/Log Pis Mean 25.271 +trainer/Log Pis Std 14.4074 +trainer/Log Pis Max 71.3496 +trainer/Log Pis Min -7.69072 +trainer/policy/mean Mean -0.0372247 +trainer/policy/mean Std 0.911121 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83039 +trainer/policy/normal/std Std 0.659585 +trainer/policy/normal/std Max 5.85427 +trainer/policy/normal/std Min 0.344257 +trainer/policy/normal/log_std Mean 1.00123 +trainer/policy/normal/log_std Std 0.316184 +trainer/policy/normal/log_std Max 1.76717 +trainer/policy/normal/log_std Min -1.06637 +eval/num steps total 916320 +eval/num paths total 920 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.280781 +eval/Actions Std 0.891885 +eval/Actions Max 0.99999 +eval/Actions Min -0.999964 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.69816 +time/logging (s) 0.00377238 +time/sampling batch (s) 0.54386 +time/saving (s) 0.00351109 +time/training (s) 7.1345 +time/epoch (s) 10.3838 +time/total (s) 9352.18 +Epoch -81 +---------------------------------- --------------- +2022-05-10 15:46:52.460968 PDT | [0] Epoch -80 finished +---------------------------------- --------------- +epoch -80 +replay_buffer/size 999033 +trainer/num train calls 921000 +trainer/Policy Loss -19.6053 +trainer/Log Pis Mean 24.2507 +trainer/Log Pis Std 13.3363 +trainer/Log Pis Max 70.4977 +trainer/Log Pis Min -7.69249 +trainer/policy/mean Mean -0.0348836 +trainer/policy/mean Std 0.906518 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.78859 +trainer/policy/normal/std Std 0.666628 +trainer/policy/normal/std Max 5.30129 +trainer/policy/normal/std Min 0.225228 +trainer/policy/normal/log_std Mean 0.984982 +trainer/policy/normal/log_std Std 0.320139 +trainer/policy/normal/log_std Max 1.66795 +trainer/policy/normal/log_std Min -1.49064 +eval/num steps total 917320 +eval/num paths total 921 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109454 +eval/Actions Std 0.918118 +eval/Actions Max 0.999985 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.84904 +time/logging (s) 0.00384984 +time/sampling batch (s) 0.291807 +time/saving (s) 0.00355608 +time/training (s) 6.61483 +time/epoch (s) 9.76308 +time/total (s) 9361.94 +Epoch -80 +---------------------------------- --------------- +2022-05-10 15:47:02.548140 PDT | [0] Epoch -79 finished +---------------------------------- --------------- +epoch -79 +replay_buffer/size 999033 +trainer/num train calls 922000 +trainer/Policy Loss -19.796 +trainer/Log Pis Mean 24.9599 +trainer/Log Pis Std 13.3376 +trainer/Log Pis Max 68.2366 +trainer/Log Pis Min -9.25444 +trainer/policy/mean Mean -0.0176395 +trainer/policy/mean Std 0.902578 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79415 +trainer/policy/normal/std Std 0.682728 +trainer/policy/normal/std Max 5.89998 +trainer/policy/normal/std Min 0.224662 +trainer/policy/normal/log_std Mean 0.982755 +trainer/policy/normal/log_std Std 0.341045 +trainer/policy/normal/log_std Max 1.77495 +trainer/policy/normal/log_std Min -1.49316 +eval/num steps total 918320 +eval/num paths total 922 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0895993 +eval/Actions Std 0.878628 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76381 +time/logging (s) 0.00388824 +time/sampling batch (s) 0.290739 +time/saving (s) 0.00351979 +time/training (s) 7.00016 +time/epoch (s) 10.0621 +time/total (s) 9372.01 +Epoch -79 +---------------------------------- --------------- +2022-05-10 15:47:12.241667 PDT | [0] Epoch -78 finished +---------------------------------- --------------- +epoch -78 +replay_buffer/size 999033 +trainer/num train calls 923000 +trainer/Policy Loss -19.429 +trainer/Log Pis Mean 24.887 +trainer/Log Pis Std 13.676 +trainer/Log Pis Max 71.4073 +trainer/Log Pis Min -5.41406 +trainer/policy/mean Mean -0.0433353 +trainer/policy/mean Std 0.905848 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.74879 +trainer/policy/normal/std Std 0.663252 +trainer/policy/normal/std Max 7.36279 +trainer/policy/normal/std Min 0.24886 +trainer/policy/normal/log_std Mean 0.968671 +trainer/policy/normal/log_std Std 0.331338 +trainer/policy/normal/log_std Max 1.99644 +trainer/policy/normal/log_std Min -1.39086 +eval/num steps total 919320 +eval/num paths total 923 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0395176 +eval/Actions Std 0.903377 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.72878 +time/logging (s) 0.00383236 +time/sampling batch (s) 0.290898 +time/saving (s) 0.00350089 +time/training (s) 6.64129 +time/epoch (s) 9.66831 +time/total (s) 9381.68 +Epoch -78 +---------------------------------- --------------- +2022-05-10 15:47:22.592070 PDT | [0] Epoch -77 finished +---------------------------------- --------------- +epoch -77 +replay_buffer/size 999033 +trainer/num train calls 924000 +trainer/Policy Loss -19.8429 +trainer/Log Pis Mean 25.5191 +trainer/Log Pis Std 13.2321 +trainer/Log Pis Max 64.3009 +trainer/Log Pis Min -6.66558 +trainer/policy/mean Mean -0.04015 +trainer/policy/mean Std 0.907262 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.77174 +trainer/policy/normal/std Std 0.656671 +trainer/policy/normal/std Max 5.66655 +trainer/policy/normal/std Min 0.243152 +trainer/policy/normal/log_std Mean 0.97893 +trainer/policy/normal/log_std Std 0.322125 +trainer/policy/normal/log_std Max 1.73458 +trainer/policy/normal/log_std Min -1.41407 +eval/num steps total 920320 +eval/num paths total 924 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.225045 +eval/Actions Std 0.741401 +eval/Actions Max 0.999995 +eval/Actions Min -0.999983 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57593 +time/logging (s) 0.00372006 +time/sampling batch (s) 0.290849 +time/saving (s) 0.00353448 +time/training (s) 7.45147 +time/epoch (s) 10.3255 +time/total (s) 9392.01 +Epoch -77 +---------------------------------- --------------- +2022-05-10 15:47:33.648527 PDT | [0] Epoch -76 finished +---------------------------------- --------------- +epoch -76 +replay_buffer/size 999033 +trainer/num train calls 925000 +trainer/Policy Loss -20.1961 +trainer/Log Pis Mean 24.8474 +trainer/Log Pis Std 13.5855 +trainer/Log Pis Max 66.0213 +trainer/Log Pis Min -3.03921 +trainer/policy/mean Mean -0.0166894 +trainer/policy/mean Std 0.909469 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.8475 +trainer/policy/normal/std Std 0.673656 +trainer/policy/normal/std Max 5.913 +trainer/policy/normal/std Min 0.299195 +trainer/policy/normal/log_std Mean 1.00559 +trainer/policy/normal/log_std Std 0.324027 +trainer/policy/normal/log_std Max 1.77715 +trainer/policy/normal/log_std Min -1.20666 +eval/num steps total 921320 +eval/num paths total 925 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.165621 +eval/Actions Std 0.940682 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43817 +time/logging (s) 0.00423487 +time/sampling batch (s) 0.782126 +time/saving (s) 0.00412729 +time/training (s) 7.80413 +time/epoch (s) 11.0328 +time/total (s) 9403.04 +Epoch -76 +---------------------------------- --------------- +2022-05-10 15:47:44.306750 PDT | [0] Epoch -75 finished +---------------------------------- --------------- +epoch -75 +replay_buffer/size 999033 +trainer/num train calls 926000 +trainer/Policy Loss -18.9306 +trainer/Log Pis Mean 24.196 +trainer/Log Pis Std 13.2916 +trainer/Log Pis Max 67.739 +trainer/Log Pis Min -6.62432 +trainer/policy/mean Mean -0.0572562 +trainer/policy/mean Std 0.903992 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83084 +trainer/policy/normal/std Std 0.658628 +trainer/policy/normal/std Max 5.8043 +trainer/policy/normal/std Min 0.245971 +trainer/policy/normal/log_std Mean 1.00016 +trainer/policy/normal/log_std Std 0.32583 +trainer/policy/normal/log_std Max 1.7586 +trainer/policy/normal/log_std Min -1.40254 +eval/num steps total 922320 +eval/num paths total 926 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.109668 +eval/Actions Std 0.908624 +eval/Actions Max 0.99999 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.46958 +time/logging (s) 0.00384428 +time/sampling batch (s) 0.280498 +time/saving (s) 0.00361267 +time/training (s) 7.87588 +time/epoch (s) 10.6334 +time/total (s) 9413.68 +Epoch -75 +---------------------------------- --------------- +2022-05-10 15:47:54.188299 PDT | [0] Epoch -74 finished +---------------------------------- --------------- +epoch -74 +replay_buffer/size 999033 +trainer/num train calls 927000 +trainer/Policy Loss -18.9292 +trainer/Log Pis Mean 24.5985 +trainer/Log Pis Std 13.3118 +trainer/Log Pis Max 65.865 +trainer/Log Pis Min -7.70766 +trainer/policy/mean Mean -0.0311622 +trainer/policy/mean Std 0.906882 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.80663 +trainer/policy/normal/std Std 0.695602 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.229941 +trainer/policy/normal/log_std Mean 0.98537 +trainer/policy/normal/log_std Std 0.351192 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.46993 +eval/num steps total 923320 +eval/num paths total 927 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.237159 +eval/Actions Std 0.943131 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58254 +time/logging (s) 0.00369244 +time/sampling batch (s) 0.278806 +time/saving (s) 0.00337979 +time/training (s) 6.98896 +time/epoch (s) 9.85737 +time/total (s) 9423.54 +Epoch -74 +---------------------------------- --------------- +2022-05-10 15:48:04.581455 PDT | [0] Epoch -73 finished +---------------------------------- --------------- +epoch -73 +replay_buffer/size 999033 +trainer/num train calls 928000 +trainer/Policy Loss -19.742 +trainer/Log Pis Mean 24.4555 +trainer/Log Pis Std 13.5463 +trainer/Log Pis Max 68.6035 +trainer/Log Pis Min -11.6051 +trainer/policy/mean Mean -0.0397856 +trainer/policy/mean Std 0.902992 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.79513 +trainer/policy/normal/std Std 0.667802 +trainer/policy/normal/std Max 5.76572 +trainer/policy/normal/std Min 0.28674 +trainer/policy/normal/log_std Mean 0.986029 +trainer/policy/normal/log_std Std 0.328116 +trainer/policy/normal/log_std Max 1.75193 +trainer/policy/normal/log_std Min -1.24918 +eval/num steps total 924320 +eval/num paths total 928 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105528 +eval/Actions Std 0.910876 +eval/Actions Max 0.999995 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.52737 +time/logging (s) 0.00370302 +time/sampling batch (s) 0.279744 +time/saving (s) 0.00346508 +time/training (s) 7.55478 +time/epoch (s) 10.3691 +time/total (s) 9433.91 +Epoch -73 +---------------------------------- --------------- +2022-05-10 15:48:14.439831 PDT | [0] Epoch -72 finished +---------------------------------- --------------- +epoch -72 +replay_buffer/size 999033 +trainer/num train calls 929000 +trainer/Policy Loss -18.0936 +trainer/Log Pis Mean 23.8135 +trainer/Log Pis Std 13.2496 +trainer/Log Pis Max 67.5305 +trainer/Log Pis Min -4.86555 +trainer/policy/mean Mean 0.00338809 +trainer/policy/mean Std 0.909048 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.83138 +trainer/policy/normal/std Std 0.687882 +trainer/policy/normal/std Max 6.21235 +trainer/policy/normal/std Min 0.256753 +trainer/policy/normal/log_std Mean 0.998812 +trainer/policy/normal/log_std Std 0.325856 +trainer/policy/normal/log_std Max 1.82654 +trainer/policy/normal/log_std Min -1.35964 +eval/num steps total 925320 +eval/num paths total 929 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0759499 +eval/Actions Std 0.897044 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54085 +time/logging (s) 0.00369164 +time/sampling batch (s) 0.280409 +time/saving (s) 0.00345528 +time/training (s) 7.00588 +time/epoch (s) 9.83429 +time/total (s) 9443.75 +Epoch -72 +---------------------------------- --------------- +2022-05-10 15:48:24.309461 PDT | [0] Epoch -71 finished +---------------------------------- --------------- +epoch -71 +replay_buffer/size 999033 +trainer/num train calls 930000 +trainer/Policy Loss -20.6346 +trainer/Log Pis Mean 25.261 +trainer/Log Pis Std 13.579 +trainer/Log Pis Max 72.8304 +trainer/Log Pis Min -5.55281 +trainer/policy/mean Mean -0.0336395 +trainer/policy/mean Std 0.910154 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.82264 +trainer/policy/normal/std Std 0.676179 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.274637 +trainer/policy/normal/log_std Mean 0.995367 +trainer/policy/normal/log_std Std 0.332222 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.29231 +eval/num steps total 926320 +eval/num paths total 930 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.318541 +eval/Actions Std 0.824764 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60788 +time/logging (s) 0.00378386 +time/sampling batch (s) 0.279605 +time/saving (s) 0.00359428 +time/training (s) 6.95083 +time/epoch (s) 9.84569 +time/total (s) 9453.6 +Epoch -71 +---------------------------------- --------------- +2022-05-10 15:48:35.074548 PDT | [0] Epoch -70 finished +---------------------------------- --------------- +epoch -70 +replay_buffer/size 999033 +trainer/num train calls 931000 +trainer/Policy Loss -19.7387 +trainer/Log Pis Mean 24.1048 +trainer/Log Pis Std 12.7337 +trainer/Log Pis Max 64.535 +trainer/Log Pis Min -9.61172 +trainer/policy/mean Mean -0.0540466 +trainer/policy/mean Std 0.90532 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.81443 +trainer/policy/normal/std Std 0.698228 +trainer/policy/normal/std Max 6.81238 +trainer/policy/normal/std Min 0.271891 +trainer/policy/normal/log_std Mean 0.990286 +trainer/policy/normal/log_std Std 0.337294 +trainer/policy/normal/log_std Max 1.91874 +trainer/policy/normal/log_std Min -1.30236 +eval/num steps total 927320 +eval/num paths total 931 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0871757 +eval/Actions Std 0.915113 +eval/Actions Max 0.999999 +eval/Actions Min -0.999984 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66314 +time/logging (s) 0.00412209 +time/sampling batch (s) 0.283534 +time/saving (s) 0.00416282 +time/training (s) 7.78604 +time/epoch (s) 10.741 +time/total (s) 9464.34 +Epoch -70 +---------------------------------- --------------- +2022-05-10 15:48:45.225676 PDT | [0] Epoch -69 finished +---------------------------------- --------------- +epoch -69 +replay_buffer/size 999033 +trainer/num train calls 932000 +trainer/Policy Loss -19.0319 +trainer/Log Pis Mean 25.016 +trainer/Log Pis Std 13.8058 +trainer/Log Pis Max 72.6914 +trainer/Log Pis Min -8.80946 +trainer/policy/mean Mean -0.035146 +trainer/policy/mean Std 0.908041 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.82737 +trainer/policy/normal/std Std 0.677453 +trainer/policy/normal/std Max 6.09271 +trainer/policy/normal/std Min 0.311106 +trainer/policy/normal/log_std Mean 0.998604 +trainer/policy/normal/log_std Std 0.321841 +trainer/policy/normal/log_std Max 1.80709 +trainer/policy/normal/log_std Min -1.16762 +eval/num steps total 928320 +eval/num paths total 932 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.054018 +eval/Actions Std 0.914074 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.44406 +time/logging (s) 0.00377921 +time/sampling batch (s) 0.280447 +time/saving (s) 0.00360141 +time/training (s) 7.39418 +time/epoch (s) 10.1261 +time/total (s) 9474.47 +Epoch -69 +---------------------------------- --------------- +2022-05-10 15:48:55.619049 PDT | [0] Epoch -68 finished +---------------------------------- --------------- +epoch -68 +replay_buffer/size 999033 +trainer/num train calls 933000 +trainer/Policy Loss -19.6831 +trainer/Log Pis Mean 24.9104 +trainer/Log Pis Std 13.594 +trainer/Log Pis Max 69.9881 +trainer/Log Pis Min -5.35678 +trainer/policy/mean Mean -0.0303213 +trainer/policy/mean Std 0.903546 +trainer/policy/mean Max 0.999977 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.73681 +trainer/policy/normal/std Std 0.689508 +trainer/policy/normal/std Max 7.27314 +trainer/policy/normal/std Min 0.267924 +trainer/policy/normal/log_std Mean 0.957899 +trainer/policy/normal/log_std Std 0.359949 +trainer/policy/normal/log_std Max 1.98419 +trainer/policy/normal/log_std Min -1.31705 +eval/num steps total 929320 +eval/num paths total 933 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0981663 +eval/Actions Std 0.914114 +eval/Actions Max 0.999992 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5786 +time/logging (s) 0.00379315 +time/sampling batch (s) 0.532393 +time/saving (s) 0.00347178 +time/training (s) 7.25095 +time/epoch (s) 10.3692 +time/total (s) 9484.84 +Epoch -68 +---------------------------------- --------------- +2022-05-10 15:49:05.089023 PDT | [0] Epoch -67 finished +---------------------------------- --------------- +epoch -67 +replay_buffer/size 999033 +trainer/num train calls 934000 +trainer/Policy Loss -19.913 +trainer/Log Pis Mean 23.8346 +trainer/Log Pis Std 13.0994 +trainer/Log Pis Max 71.5076 +trainer/Log Pis Min -9.53119 +trainer/policy/mean Mean -0.0453809 +trainer/policy/mean Std 0.903735 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.70634 +trainer/policy/normal/std Std 0.684101 +trainer/policy/normal/std Max 5.43291 +trainer/policy/normal/std Min 0.255421 +trainer/policy/normal/log_std Mean 0.945873 +trainer/policy/normal/log_std Std 0.363066 +trainer/policy/normal/log_std Max 1.69247 +trainer/policy/normal/log_std Min -1.36484 +eval/num steps total 930320 +eval/num paths total 934 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.128837 +eval/Actions Std 0.907792 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57907 +time/logging (s) 0.00368165 +time/sampling batch (s) 0.52658 +time/saving (s) 0.00357113 +time/training (s) 6.33321 +time/epoch (s) 9.44611 +time/total (s) 9494.29 +Epoch -67 +---------------------------------- --------------- +2022-05-10 15:49:15.178475 PDT | [0] Epoch -66 finished +---------------------------------- --------------- +epoch -66 +replay_buffer/size 999033 +trainer/num train calls 935000 +trainer/Policy Loss -20.0521 +trainer/Log Pis Mean 24.4477 +trainer/Log Pis Std 13.2727 +trainer/Log Pis Max 66.0325 +trainer/Log Pis Min -6.91109 +trainer/policy/mean Mean -0.0199167 +trainer/policy/mean Std 0.90378 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.85461 +trainer/policy/normal/std Std 0.705086 +trainer/policy/normal/std Max 6.50824 +trainer/policy/normal/std Min 0.279886 +trainer/policy/normal/log_std Mean 1.00419 +trainer/policy/normal/log_std Std 0.33981 +trainer/policy/normal/log_std Max 1.87307 +trainer/policy/normal/log_std Min -1.27337 +eval/num steps total 931320 +eval/num paths total 935 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.123834 +eval/Actions Std 0.904476 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40462 +time/logging (s) 0.00372722 +time/sampling batch (s) 0.528705 +time/saving (s) 0.00345391 +time/training (s) 7.12497 +time/epoch (s) 10.0655 +time/total (s) 9504.36 +Epoch -66 +---------------------------------- --------------- +2022-05-10 15:49:25.134230 PDT | [0] Epoch -65 finished +---------------------------------- --------------- +epoch -65 +replay_buffer/size 999033 +trainer/num train calls 936000 +trainer/Policy Loss -19.9945 +trainer/Log Pis Mean 24.6606 +trainer/Log Pis Std 12.7328 +trainer/Log Pis Max 62.711 +trainer/Log Pis Min -4.73034 +trainer/policy/mean Mean -0.049404 +trainer/policy/mean Std 0.908374 +trainer/policy/mean Max 0.999978 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.76897 +trainer/policy/normal/std Std 0.666481 +trainer/policy/normal/std Max 5.27414 +trainer/policy/normal/std Min 0.237076 +trainer/policy/normal/log_std Mean 0.975297 +trainer/policy/normal/log_std Std 0.335153 +trainer/policy/normal/log_std Max 1.66282 +trainer/policy/normal/log_std Min -1.43937 +eval/num steps total 932320 +eval/num paths total 936 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.016386 +eval/Actions Std 0.911135 +eval/Actions Max 0.999994 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.73044 +time/logging (s) 0.0037 +time/sampling batch (s) 0.531216 +time/saving (s) 0.00347434 +time/training (s) 6.66275 +time/epoch (s) 9.93158 +time/total (s) 9514.29 +Epoch -65 +---------------------------------- --------------- +2022-05-10 15:49:35.284104 PDT | [0] Epoch -64 finished +---------------------------------- --------------- +epoch -64 +replay_buffer/size 999033 +trainer/num train calls 937000 +trainer/Policy Loss -18.7622 +trainer/Log Pis Mean 24.1068 +trainer/Log Pis Std 14.0867 +trainer/Log Pis Max 71.1984 +trainer/Log Pis Min -10.8242 +trainer/policy/mean Mean -0.0409328 +trainer/policy/mean Std 0.904209 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76389 +trainer/policy/normal/std Std 0.675211 +trainer/policy/normal/std Max 5.98521 +trainer/policy/normal/std Min 0.27145 +trainer/policy/normal/log_std Mean 0.972916 +trainer/policy/normal/log_std Std 0.335915 +trainer/policy/normal/log_std Max 1.78929 +trainer/policy/normal/log_std Min -1.30398 +eval/num steps total 933320 +eval/num paths total 937 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0787566 +eval/Actions Std 0.912956 +eval/Actions Max 0.999998 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.39104 +time/logging (s) 0.00371349 +time/sampling batch (s) 0.284267 +time/saving (s) 0.0034608 +time/training (s) 7.44292 +time/epoch (s) 10.1254 +time/total (s) 9524.42 +Epoch -64 +---------------------------------- --------------- +2022-05-10 15:49:46.615794 PDT | [0] Epoch -63 finished +---------------------------------- --------------- +epoch -63 +replay_buffer/size 999033 +trainer/num train calls 938000 +trainer/Policy Loss -20.2971 +trainer/Log Pis Mean 25.2456 +trainer/Log Pis Std 13.5683 +trainer/Log Pis Max 74.1846 +trainer/Log Pis Min -6.68819 +trainer/policy/mean Mean -0.0311395 +trainer/policy/mean Std 0.909714 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82012 +trainer/policy/normal/std Std 0.645628 +trainer/policy/normal/std Max 6.03844 +trainer/policy/normal/std Min 0.31144 +trainer/policy/normal/log_std Mean 0.999306 +trainer/policy/normal/log_std Std 0.308726 +trainer/policy/normal/log_std Max 1.79815 +trainer/policy/normal/log_std Min -1.16655 +eval/num steps total 934320 +eval/num paths total 938 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.249683 +eval/Actions Std 0.832349 +eval/Actions Max 0.999998 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47563 +time/logging (s) 0.00397249 +time/sampling batch (s) 0.289371 +time/saving (s) 0.00411229 +time/training (s) 8.53419 +time/epoch (s) 11.3073 +time/total (s) 9535.73 +Epoch -63 +---------------------------------- --------------- +2022-05-10 15:49:56.629607 PDT | [0] Epoch -62 finished +---------------------------------- --------------- +epoch -62 +replay_buffer/size 999033 +trainer/num train calls 939000 +trainer/Policy Loss -20.5832 +trainer/Log Pis Mean 25.7505 +trainer/Log Pis Std 13.7188 +trainer/Log Pis Max 71.6347 +trainer/Log Pis Min -4.91441 +trainer/policy/mean Mean -0.0340519 +trainer/policy/mean Std 0.906802 +trainer/policy/mean Max 0.999995 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83127 +trainer/policy/normal/std Std 0.699934 +trainer/policy/normal/std Max 6.00528 +trainer/policy/normal/std Min 0.269073 +trainer/policy/normal/log_std Mean 0.995654 +trainer/policy/normal/log_std Std 0.341366 +trainer/policy/normal/log_std Max 1.79264 +trainer/policy/normal/log_std Min -1.31277 +eval/num steps total 935320 +eval/num paths total 939 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0680796 +eval/Actions Std 0.832616 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62965 +time/logging (s) 0.00397597 +time/sampling batch (s) 0.280148 +time/saving (s) 0.00359861 +time/training (s) 7.07143 +time/epoch (s) 9.9888 +time/total (s) 9545.72 +Epoch -62 +---------------------------------- --------------- +2022-05-10 15:50:06.375140 PDT | [0] Epoch -61 finished +---------------------------------- --------------- +epoch -61 +replay_buffer/size 999033 +trainer/num train calls 940000 +trainer/Policy Loss -19.7881 +trainer/Log Pis Mean 23.4252 +trainer/Log Pis Std 13.671 +trainer/Log Pis Max 61.5275 +trainer/Log Pis Min -13.5846 +trainer/policy/mean Mean -0.0399554 +trainer/policy/mean Std 0.900213 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.77194 +trainer/policy/normal/std Std 0.693628 +trainer/policy/normal/std Max 6.4703 +trainer/policy/normal/std Min 0.285803 +trainer/policy/normal/log_std Mean 0.974015 +trainer/policy/normal/log_std Std 0.342513 +trainer/policy/normal/log_std Max 1.86722 +trainer/policy/normal/log_std Min -1.25245 +eval/num steps total 936320 +eval/num paths total 940 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.072299 +eval/Actions Std 0.903518 +eval/Actions Max 1 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66823 +time/logging (s) 0.00384196 +time/sampling batch (s) 0.279216 +time/saving (s) 0.00347161 +time/training (s) 6.76633 +time/epoch (s) 9.72109 +time/total (s) 9555.45 +Epoch -61 +---------------------------------- --------------- +2022-05-10 15:50:15.569615 PDT | [0] Epoch -60 finished +---------------------------------- --------------- +epoch -60 +replay_buffer/size 999033 +trainer/num train calls 941000 +trainer/Policy Loss -19.6941 +trainer/Log Pis Mean 23.8382 +trainer/Log Pis Std 13.6674 +trainer/Log Pis Max 76.0205 +trainer/Log Pis Min -5.79548 +trainer/policy/mean Mean -0.050002 +trainer/policy/mean Std 0.905053 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999989 +trainer/policy/normal/std Mean 2.83695 +trainer/policy/normal/std Std 0.675892 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.321089 +trainer/policy/normal/log_std Mean 1.00246 +trainer/policy/normal/log_std Std 0.319215 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.13604 +eval/num steps total 937320 +eval/num paths total 941 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.09015 +eval/Actions Std 0.919485 +eval/Actions Max 0.999996 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.43481 +time/logging (s) 0.00378506 +time/sampling batch (s) 0.279108 +time/saving (s) 0.00346915 +time/training (s) 6.44912 +time/epoch (s) 9.17029 +time/total (s) 9564.62 +Epoch -60 +---------------------------------- --------------- +2022-05-10 15:50:24.283304 PDT | [0] Epoch -59 finished +---------------------------------- --------------- +epoch -59 +replay_buffer/size 999033 +trainer/num train calls 942000 +trainer/Policy Loss -20.1106 +trainer/Log Pis Mean 23.9996 +trainer/Log Pis Std 13.5208 +trainer/Log Pis Max 69.6971 +trainer/Log Pis Min -11.5614 +trainer/policy/mean Mean -0.0184727 +trainer/policy/mean Std 0.907506 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80698 +trainer/policy/normal/std Std 0.675175 +trainer/policy/normal/std Max 5.79036 +trainer/policy/normal/std Min 0.283591 +trainer/policy/normal/log_std Mean 0.989456 +trainer/policy/normal/log_std Std 0.33138 +trainer/policy/normal/log_std Max 1.7562 +trainer/policy/normal/log_std Min -1.26022 +eval/num steps total 938320 +eval/num paths total 942 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141671 +eval/Actions Std 0.920804 +eval/Actions Max 0.999996 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.26232 +time/logging (s) 0.00377823 +time/sampling batch (s) 0.279932 +time/saving (s) 0.00342489 +time/training (s) 6.13997 +time/epoch (s) 8.68942 +time/total (s) 9573.31 +Epoch -59 +---------------------------------- --------------- +2022-05-10 15:50:34.240429 PDT | [0] Epoch -58 finished +---------------------------------- --------------- +epoch -58 +replay_buffer/size 999033 +trainer/num train calls 943000 +trainer/Policy Loss -19.3597 +trainer/Log Pis Mean 24.2614 +trainer/Log Pis Std 12.895 +trainer/Log Pis Max 62.757 +trainer/Log Pis Min -11.4766 +trainer/policy/mean Mean -0.0211311 +trainer/policy/mean Std 0.90491 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.77454 +trainer/policy/normal/std Std 0.650323 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.294933 +trainer/policy/normal/log_std Mean 0.981054 +trainer/policy/normal/log_std Std 0.317289 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.22101 +eval/num steps total 939320 +eval/num paths total 943 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0898896 +eval/Actions Std 0.916982 +eval/Actions Max 0.99999 +eval/Actions Min -0.999989 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60904 +time/logging (s) 0.00374293 +time/sampling batch (s) 0.2783 +time/saving (s) 0.00344254 +time/training (s) 7.0385 +time/epoch (s) 9.93302 +time/total (s) 9583.25 +Epoch -58 +---------------------------------- --------------- +2022-05-10 15:50:44.155195 PDT | [0] Epoch -57 finished +---------------------------------- --------------- +epoch -57 +replay_buffer/size 999033 +trainer/num train calls 944000 +trainer/Policy Loss -20.0464 +trainer/Log Pis Mean 24.8855 +trainer/Log Pis Std 13.0678 +trainer/Log Pis Max 61.149 +trainer/Log Pis Min -9.97343 +trainer/policy/mean Mean -0.0428064 +trainer/policy/mean Std 0.902125 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.86585 +trainer/policy/normal/std Std 0.678607 +trainer/policy/normal/std Max 6.94301 +trainer/policy/normal/std Min 0.296404 +trainer/policy/normal/log_std Mean 1.01282 +trainer/policy/normal/log_std Std 0.319008 +trainer/policy/normal/log_std Max 1.93774 +trainer/policy/normal/log_std Min -1.21603 +eval/num steps total 940320 +eval/num paths total 944 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0133906 +eval/Actions Std 0.96376 +eval/Actions Max 0.999974 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.54845 +time/logging (s) 0.0043923 +time/sampling batch (s) 0.280964 +time/saving (s) 0.00422074 +time/training (s) 7.05315 +time/epoch (s) 9.89118 +time/total (s) 9593.14 +Epoch -57 +---------------------------------- --------------- +2022-05-10 15:50:53.989301 PDT | [0] Epoch -56 finished +---------------------------------- --------------- +epoch -56 +replay_buffer/size 999033 +trainer/num train calls 945000 +trainer/Policy Loss -18.4633 +trainer/Log Pis Mean 23.5003 +trainer/Log Pis Std 13.2233 +trainer/Log Pis Max 74.5697 +trainer/Log Pis Min -5.86044 +trainer/policy/mean Mean -0.0352772 +trainer/policy/mean Std 0.907283 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.79056 +trainer/policy/normal/std Std 0.665807 +trainer/policy/normal/std Max 6.54256 +trainer/policy/normal/std Min 0.320601 +trainer/policy/normal/log_std Mean 0.984344 +trainer/policy/normal/log_std Std 0.328945 +trainer/policy/normal/log_std Max 1.87833 +trainer/policy/normal/log_std Min -1.13756 +eval/num steps total 941320 +eval/num paths total 945 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0268201 +eval/Actions Std 0.907909 +eval/Actions Max 0.999997 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63733 +time/logging (s) 0.00370337 +time/sampling batch (s) 0.533702 +time/saving (s) 0.00345899 +time/training (s) 6.62949 +time/epoch (s) 9.80768 +time/total (s) 9602.96 +Epoch -56 +---------------------------------- --------------- +2022-05-10 15:51:04.851680 PDT | [0] Epoch -55 finished +---------------------------------- --------------- +epoch -55 +replay_buffer/size 999033 +trainer/num train calls 946000 +trainer/Policy Loss -20.1645 +trainer/Log Pis Mean 23.6837 +trainer/Log Pis Std 13.9751 +trainer/Log Pis Max 66.0027 +trainer/Log Pis Min -11.6777 +trainer/policy/mean Mean -0.0383983 +trainer/policy/mean Std 0.906246 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999983 +trainer/policy/normal/std Mean 2.74557 +trainer/policy/normal/std Std 0.657432 +trainer/policy/normal/std Max 5.18854 +trainer/policy/normal/std Min 0.210186 +trainer/policy/normal/log_std Mean 0.966901 +trainer/policy/normal/log_std Std 0.33609 +trainer/policy/normal/log_std Max 1.64645 +trainer/policy/normal/log_std Min -1.55976 +eval/num steps total 942320 +eval/num paths total 946 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0329893 +eval/Actions Std 0.916082 +eval/Actions Max 0.999993 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.47128 +time/logging (s) 0.00419425 +time/sampling batch (s) 0.28698 +time/saving (s) 0.00406043 +time/training (s) 8.07175 +time/epoch (s) 10.8383 +time/total (s) 9613.8 +Epoch -55 +---------------------------------- --------------- +2022-05-10 15:51:14.699618 PDT | [0] Epoch -54 finished +---------------------------------- --------------- +epoch -54 +replay_buffer/size 999033 +trainer/num train calls 947000 +trainer/Policy Loss -19.2263 +trainer/Log Pis Mean 24.9063 +trainer/Log Pis Std 13.4926 +trainer/Log Pis Max 67.0698 +trainer/Log Pis Min -7.97591 +trainer/policy/mean Mean -0.0371914 +trainer/policy/mean Std 0.910046 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999981 +trainer/policy/normal/std Mean 2.82357 +trainer/policy/normal/std Std 0.671563 +trainer/policy/normal/std Max 6.48556 +trainer/policy/normal/std Min 0.265102 +trainer/policy/normal/log_std Mean 0.995408 +trainer/policy/normal/log_std Std 0.334383 +trainer/policy/normal/log_std Max 1.86958 +trainer/policy/normal/log_std Min -1.32764 +eval/num steps total 943320 +eval/num paths total 947 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.141402 +eval/Actions Std 0.898324 +eval/Actions Max 0.99997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50742 +time/logging (s) 0.00447409 +time/sampling batch (s) 0.284437 +time/saving (s) 0.00411474 +time/training (s) 7.02283 +time/epoch (s) 9.82328 +time/total (s) 9623.62 +Epoch -54 +---------------------------------- --------------- +2022-05-10 15:51:24.298652 PDT | [0] Epoch -53 finished +---------------------------------- --------------- +epoch -53 +replay_buffer/size 999033 +trainer/num train calls 948000 +trainer/Policy Loss -19.7581 +trainer/Log Pis Mean 25.048 +trainer/Log Pis Std 13.3284 +trainer/Log Pis Max 67.6258 +trainer/Log Pis Min -9.81606 +trainer/policy/mean Mean -0.0273061 +trainer/policy/mean Std 0.908728 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.77523 +trainer/policy/normal/std Std 0.67841 +trainer/policy/normal/std Max 6.36537 +trainer/policy/normal/std Min 0.239748 +trainer/policy/normal/log_std Mean 0.975906 +trainer/policy/normal/log_std Std 0.343501 +trainer/policy/normal/log_std Max 1.85087 +trainer/policy/normal/log_std Min -1.42817 +eval/num steps total 944320 +eval/num paths total 948 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.479866 +eval/Actions Std 0.767438 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.77453 +time/logging (s) 0.0038207 +time/sampling batch (s) 0.282952 +time/saving (s) 0.00346366 +time/training (s) 6.50834 +time/epoch (s) 9.57311 +time/total (s) 9633.2 +Epoch -53 +---------------------------------- --------------- +2022-05-10 15:51:33.973436 PDT | [0] Epoch -52 finished +---------------------------------- --------------- +epoch -52 +replay_buffer/size 999033 +trainer/num train calls 949000 +trainer/Policy Loss -19.268 +trainer/Log Pis Mean 25.2755 +trainer/Log Pis Std 14.6604 +trainer/Log Pis Max 78.1837 +trainer/Log Pis Min -8.29071 +trainer/policy/mean Mean -0.045492 +trainer/policy/mean Std 0.909883 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.80693 +trainer/policy/normal/std Std 0.683078 +trainer/policy/normal/std Max 5.64601 +trainer/policy/normal/std Min 0.250173 +trainer/policy/normal/log_std Mean 0.988179 +trainer/policy/normal/log_std Std 0.337939 +trainer/policy/normal/log_std Max 1.73095 +trainer/policy/normal/log_std Min -1.3856 +eval/num steps total 945320 +eval/num paths total 949 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.105818 +eval/Actions Std 0.919716 +eval/Actions Max 0.999998 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55651 +time/logging (s) 0.00378405 +time/sampling batch (s) 0.284011 +time/saving (s) 0.00344491 +time/training (s) 6.80267 +time/epoch (s) 9.65041 +time/total (s) 9642.85 +Epoch -52 +---------------------------------- --------------- +2022-05-10 15:51:43.307564 PDT | [0] Epoch -51 finished +---------------------------------- --------------- +epoch -51 +replay_buffer/size 999033 +trainer/num train calls 950000 +trainer/Policy Loss -18.1761 +trainer/Log Pis Mean 24.9489 +trainer/Log Pis Std 13.258 +trainer/Log Pis Max 71.5267 +trainer/Log Pis Min -6.18773 +trainer/policy/mean Mean -0.0173839 +trainer/policy/mean Std 0.901711 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.85127 +trainer/policy/normal/std Std 0.696784 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.25411 +trainer/policy/normal/log_std Mean 1.00425 +trainer/policy/normal/log_std Std 0.336003 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.36999 +eval/num steps total 946320 +eval/num paths total 950 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.217169 +eval/Actions Std 0.920726 +eval/Actions Max 0.999998 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60105 +time/logging (s) 0.00369164 +time/sampling batch (s) 0.280048 +time/saving (s) 0.00341086 +time/training (s) 6.42157 +time/epoch (s) 9.30977 +time/total (s) 9652.16 +Epoch -51 +---------------------------------- --------------- +2022-05-10 15:51:52.825315 PDT | [0] Epoch -50 finished +---------------------------------- --------------- +epoch -50 +replay_buffer/size 999033 +trainer/num train calls 951000 +trainer/Policy Loss -20.9066 +trainer/Log Pis Mean 24.0535 +trainer/Log Pis Std 13.3974 +trainer/Log Pis Max 64.9777 +trainer/Log Pis Min -6.47147 +trainer/policy/mean Mean -0.0367015 +trainer/policy/mean Std 0.909908 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.77196 +trainer/policy/normal/std Std 0.67216 +trainer/policy/normal/std Max 5.64834 +trainer/policy/normal/std Min 0.300467 +trainer/policy/normal/log_std Mean 0.976991 +trainer/policy/normal/log_std Std 0.329846 +trainer/policy/normal/log_std Max 1.73136 +trainer/policy/normal/log_std Min -1.20242 +eval/num steps total 947320 +eval/num paths total 951 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0775004 +eval/Actions Std 0.842812 +eval/Actions Max 0.999992 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5711 +time/logging (s) 0.00368792 +time/sampling batch (s) 0.282899 +time/saving (s) 0.00346745 +time/training (s) 6.63226 +time/epoch (s) 9.49341 +time/total (s) 9661.66 +Epoch -50 +---------------------------------- --------------- +2022-05-10 15:52:02.587539 PDT | [0] Epoch -49 finished +---------------------------------- --------------- +epoch -49 +replay_buffer/size 999033 +trainer/num train calls 952000 +trainer/Policy Loss -19.6882 +trainer/Log Pis Mean 25.3108 +trainer/Log Pis Std 13.7435 +trainer/Log Pis Max 65.6662 +trainer/Log Pis Min -8.4132 +trainer/policy/mean Mean -0.0527114 +trainer/policy/mean Std 0.908194 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.81189 +trainer/policy/normal/std Std 0.661443 +trainer/policy/normal/std Max 6.28505 +trainer/policy/normal/std Min 0.327958 +trainer/policy/normal/log_std Mean 0.993566 +trainer/policy/normal/log_std Std 0.321811 +trainer/policy/normal/log_std Max 1.83817 +trainer/policy/normal/log_std Min -1.11487 +eval/num steps total 948320 +eval/num paths total 952 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0517337 +eval/Actions Std 0.913979 +eval/Actions Max 0.999989 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.34694 +time/logging (s) 0.00366225 +time/sampling batch (s) 0.532231 +time/saving (s) 0.00342349 +time/training (s) 6.85157 +time/epoch (s) 9.73783 +time/total (s) 9671.4 +Epoch -49 +---------------------------------- --------------- +2022-05-10 15:52:11.940186 PDT | [0] Epoch -48 finished +---------------------------------- --------------- +epoch -48 +replay_buffer/size 999033 +trainer/num train calls 953000 +trainer/Policy Loss -19.9844 +trainer/Log Pis Mean 23.4211 +trainer/Log Pis Std 13.377 +trainer/Log Pis Max 71.5688 +trainer/Log Pis Min -11.382 +trainer/policy/mean Mean -0.036681 +trainer/policy/mean Std 0.904381 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81425 +trainer/policy/normal/std Std 0.669804 +trainer/policy/normal/std Max 5.9131 +trainer/policy/normal/std Min 0.303946 +trainer/policy/normal/log_std Mean 0.993122 +trainer/policy/normal/log_std Std 0.328737 +trainer/policy/normal/log_std Max 1.77717 +trainer/policy/normal/log_std Min -1.19091 +eval/num steps total 949320 +eval/num paths total 953 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.198697 +eval/Actions Std 0.886216 +eval/Actions Max 0.999991 +eval/Actions Min -0.999962 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62568 +time/logging (s) 0.00370466 +time/sampling batch (s) 0.281417 +time/saving (s) 0.00347012 +time/training (s) 6.41426 +time/epoch (s) 9.32853 +time/total (s) 9680.73 +Epoch -48 +---------------------------------- --------------- +2022-05-10 15:52:21.236291 PDT | [0] Epoch -47 finished +---------------------------------- --------------- +epoch -47 +replay_buffer/size 999033 +trainer/num train calls 954000 +trainer/Policy Loss -19.8253 +trainer/Log Pis Mean 25.018 +trainer/Log Pis Std 13.6766 +trainer/Log Pis Max 72.3183 +trainer/Log Pis Min -6.20943 +trainer/policy/mean Mean -0.0342277 +trainer/policy/mean Std 0.908026 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79225 +trainer/policy/normal/std Std 0.690994 +trainer/policy/normal/std Max 5.7821 +trainer/policy/normal/std Min 0.275485 +trainer/policy/normal/log_std Mean 0.980595 +trainer/policy/normal/log_std Std 0.348267 +trainer/policy/normal/log_std Max 1.75477 +trainer/policy/normal/log_std Min -1.28922 +eval/num steps total 950320 +eval/num paths total 954 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0647224 +eval/Actions Std 0.918183 +eval/Actions Max 0.999999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.78026 +time/logging (s) 0.00413605 +time/sampling batch (s) 0.281043 +time/saving (s) 0.00412903 +time/training (s) 6.20253 +time/epoch (s) 9.2721 +time/total (s) 9690.01 +Epoch -47 +---------------------------------- --------------- +2022-05-10 15:52:31.074394 PDT | [0] Epoch -46 finished +---------------------------------- --------------- +epoch -46 +replay_buffer/size 999033 +trainer/num train calls 955000 +trainer/Policy Loss -19.9676 +trainer/Log Pis Mean 25.3262 +trainer/Log Pis Std 13.4014 +trainer/Log Pis Max 73.6248 +trainer/Log Pis Min -4.694 +trainer/policy/mean Mean -0.0409013 +trainer/policy/mean Std 0.910543 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.78287 +trainer/policy/normal/std Std 0.643154 +trainer/policy/normal/std Max 5.49953 +trainer/policy/normal/std Min 0.267922 +trainer/policy/normal/log_std Mean 0.985288 +trainer/policy/normal/log_std Std 0.311232 +trainer/policy/normal/log_std Max 1.70466 +trainer/policy/normal/log_std Min -1.31706 +eval/num steps total 951320 +eval/num paths total 955 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.121936 +eval/Actions Std 0.893819 +eval/Actions Max 0.999997 +eval/Actions Min -0.99999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5521 +time/logging (s) 0.00384078 +time/sampling batch (s) 0.282764 +time/saving (s) 0.00354649 +time/training (s) 6.97056 +time/epoch (s) 9.81281 +time/total (s) 9699.82 +Epoch -46 +---------------------------------- --------------- +2022-05-10 15:52:41.285177 PDT | [0] Epoch -45 finished +---------------------------------- --------------- +epoch -45 +replay_buffer/size 999033 +trainer/num train calls 956000 +trainer/Policy Loss -19.543 +trainer/Log Pis Mean 25.1724 +trainer/Log Pis Std 13.8233 +trainer/Log Pis Max 70.0692 +trainer/Log Pis Min -5.48181 +trainer/policy/mean Mean -0.0362005 +trainer/policy/mean Std 0.904748 +trainer/policy/mean Max 0.999969 +trainer/policy/mean Min -0.999991 +trainer/policy/normal/std Mean 2.8285 +trainer/policy/normal/std Std 0.69798 +trainer/policy/normal/std Max 6.32109 +trainer/policy/normal/std Min 0.304429 +trainer/policy/normal/log_std Mean 0.994835 +trainer/policy/normal/log_std Std 0.340869 +trainer/policy/normal/log_std Max 1.84389 +trainer/policy/normal/log_std Min -1.18932 +eval/num steps total 952320 +eval/num paths total 956 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0480963 +eval/Actions Std 0.871471 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.68934 +time/logging (s) 0.00373769 +time/sampling batch (s) 0.281122 +time/saving (s) 0.00358153 +time/training (s) 7.20836 +time/epoch (s) 10.1861 +time/total (s) 9710.01 +Epoch -45 +---------------------------------- --------------- +2022-05-10 15:52:51.264820 PDT | [0] Epoch -44 finished +---------------------------------- --------------- +epoch -44 +replay_buffer/size 999033 +trainer/num train calls 957000 +trainer/Policy Loss -20.5491 +trainer/Log Pis Mean 24.0278 +trainer/Log Pis Std 13.1752 +trainer/Log Pis Max 77.1979 +trainer/Log Pis Min -8.07135 +trainer/policy/mean Mean -0.0270929 +trainer/policy/mean Std 0.905938 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76746 +trainer/policy/normal/std Std 0.674884 +trainer/policy/normal/std Max 6.0744 +trainer/policy/normal/std Min 0.278056 +trainer/policy/normal/log_std Mean 0.973657 +trainer/policy/normal/log_std Std 0.339578 +trainer/policy/normal/log_std Max 1.80408 +trainer/policy/normal/log_std Min -1.27993 +eval/num steps total 953320 +eval/num paths total 957 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0611607 +eval/Actions Std 0.869659 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53137 +time/logging (s) 0.0037973 +time/sampling batch (s) 0.280898 +time/saving (s) 0.00343624 +time/training (s) 7.13578 +time/epoch (s) 9.95528 +time/total (s) 9719.97 +Epoch -44 +---------------------------------- --------------- +2022-05-10 15:53:01.350320 PDT | [0] Epoch -43 finished +---------------------------------- --------------- +epoch -43 +replay_buffer/size 999033 +trainer/num train calls 958000 +trainer/Policy Loss -19.2597 +trainer/Log Pis Mean 25.1891 +trainer/Log Pis Std 12.9875 +trainer/Log Pis Max 78.2026 +trainer/Log Pis Min -4.41103 +trainer/policy/mean Mean -0.0260856 +trainer/policy/mean Std 0.909905 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77079 +trainer/policy/normal/std Std 0.685124 +trainer/policy/normal/std Max 6.19033 +trainer/policy/normal/std Min 0.193116 +trainer/policy/normal/log_std Mean 0.972235 +trainer/policy/normal/log_std Std 0.354174 +trainer/policy/normal/log_std Max 1.82299 +trainer/policy/normal/log_std Min -1.64447 +eval/num steps total 954320 +eval/num paths total 958 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0162744 +eval/Actions Std 0.890047 +eval/Actions Max 0.999991 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.55858 +time/logging (s) 0.00375441 +time/sampling batch (s) 0.28111 +time/saving (s) 0.00344913 +time/training (s) 7.21415 +time/epoch (s) 10.061 +time/total (s) 9730.04 +Epoch -43 +---------------------------------- --------------- +2022-05-10 15:53:11.131680 PDT | [0] Epoch -42 finished +---------------------------------- --------------- +epoch -42 +replay_buffer/size 999033 +trainer/num train calls 959000 +trainer/Policy Loss -18.9991 +trainer/Log Pis Mean 24.8473 +trainer/Log Pis Std 13.1388 +trainer/Log Pis Max 74.2775 +trainer/Log Pis Min -12.3009 +trainer/policy/mean Mean -0.0393958 +trainer/policy/mean Std 0.908802 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.81651 +trainer/policy/normal/std Std 0.704201 +trainer/policy/normal/std Max 6.4807 +trainer/policy/normal/std Min 0.280431 +trainer/policy/normal/log_std Mean 0.988294 +trainer/policy/normal/log_std Std 0.35177 +trainer/policy/normal/log_std Max 1.86883 +trainer/policy/normal/log_std Min -1.27143 +eval/num steps total 955320 +eval/num paths total 959 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0587396 +eval/Actions Std 0.918225 +eval/Actions Max 0.999996 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53114 +time/logging (s) 0.00378363 +time/sampling batch (s) 0.281328 +time/saving (s) 0.00348608 +time/training (s) 6.93728 +time/epoch (s) 9.75702 +time/total (s) 9739.8 +Epoch -42 +---------------------------------- --------------- +2022-05-10 15:53:22.015024 PDT | [0] Epoch -41 finished +---------------------------------- --------------- +epoch -41 +replay_buffer/size 999033 +trainer/num train calls 960000 +trainer/Policy Loss -19.9928 +trainer/Log Pis Mean 24.2633 +trainer/Log Pis Std 13.1177 +trainer/Log Pis Max 75.0178 +trainer/Log Pis Min -13.0425 +trainer/policy/mean Mean -0.0483522 +trainer/policy/mean Std 0.907613 +trainer/policy/mean Max 0.999985 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80642 +trainer/policy/normal/std Std 0.675534 +trainer/policy/normal/std Max 6.08038 +trainer/policy/normal/std Min 0.243489 +trainer/policy/normal/log_std Mean 0.988697 +trainer/policy/normal/log_std Std 0.336138 +trainer/policy/normal/log_std Max 1.80507 +trainer/policy/normal/log_std Min -1.41268 +eval/num steps total 956320 +eval/num paths total 960 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.158806 +eval/Actions Std 0.938289 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67424 +time/logging (s) 0.00370358 +time/sampling batch (s) 0.281236 +time/saving (s) 0.00349324 +time/training (s) 7.8959 +time/epoch (s) 10.8586 +time/total (s) 9750.66 +Epoch -41 +---------------------------------- --------------- +2022-05-10 15:53:32.036394 PDT | [0] Epoch -40 finished +---------------------------------- --------------- +epoch -40 +replay_buffer/size 999033 +trainer/num train calls 961000 +trainer/Policy Loss -19.0021 +trainer/Log Pis Mean 24.6117 +trainer/Log Pis Std 13.3378 +trainer/Log Pis Max 73.8317 +trainer/Log Pis Min -5.28861 +trainer/policy/mean Mean -0.0362485 +trainer/policy/mean Std 0.904857 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.8239 +trainer/policy/normal/std Std 0.694271 +trainer/policy/normal/std Max 6.166 +trainer/policy/normal/std Min 0.259247 +trainer/policy/normal/log_std Mean 0.992896 +trainer/policy/normal/log_std Std 0.34368 +trainer/policy/normal/log_std Max 1.81905 +trainer/policy/normal/log_std Min -1.34997 +eval/num steps total 957320 +eval/num paths total 961 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.372936 +eval/Actions Std 0.818821 +eval/Actions Max 0.999991 +eval/Actions Min -0.99996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.66735 +time/logging (s) 0.00417212 +time/sampling batch (s) 0.532623 +time/saving (s) 0.00409033 +time/training (s) 6.78924 +time/epoch (s) 9.99748 +time/total (s) 9760.66 +Epoch -40 +---------------------------------- --------------- +2022-05-10 15:53:42.378313 PDT | [0] Epoch -39 finished +---------------------------------- --------------- +epoch -39 +replay_buffer/size 999033 +trainer/num train calls 962000 +trainer/Policy Loss -19.4359 +trainer/Log Pis Mean 23.6238 +trainer/Log Pis Std 13.3879 +trainer/Log Pis Max 63.3583 +trainer/Log Pis Min -8.49506 +trainer/policy/mean Mean -0.0503092 +trainer/policy/mean Std 0.903905 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.74795 +trainer/policy/normal/std Std 0.650353 +trainer/policy/normal/std Max 5.39185 +trainer/policy/normal/std Min 0.293351 +trainer/policy/normal/log_std Mean 0.970102 +trainer/policy/normal/log_std Std 0.322851 +trainer/policy/normal/log_std Max 1.68489 +trainer/policy/normal/log_std Min -1.22638 +eval/num steps total 958320 +eval/num paths total 962 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0606278 +eval/Actions Std 0.901236 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67346 +time/logging (s) 0.00383043 +time/sampling batch (s) 0.282852 +time/saving (s) 0.00368406 +time/training (s) 7.35262 +time/epoch (s) 10.3165 +time/total (s) 9770.98 +Epoch -39 +---------------------------------- --------------- +2022-05-10 15:53:52.260253 PDT | [0] Epoch -38 finished +---------------------------------- --------------- +epoch -38 +replay_buffer/size 999033 +trainer/num train calls 963000 +trainer/Policy Loss -19.7018 +trainer/Log Pis Mean 25.1552 +trainer/Log Pis Std 13.2517 +trainer/Log Pis Max 73.636 +trainer/Log Pis Min -8.7006 +trainer/policy/mean Mean -0.0401131 +trainer/policy/mean Std 0.905193 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.82194 +trainer/policy/normal/std Std 0.68566 +trainer/policy/normal/std Max 5.73318 +trainer/policy/normal/std Min 0.295121 +trainer/policy/normal/log_std Mean 0.993537 +trainer/policy/normal/log_std Std 0.338097 +trainer/policy/normal/log_std Max 1.74627 +trainer/policy/normal/log_std Min -1.22037 +eval/num steps total 959320 +eval/num paths total 963 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.395225 +eval/Actions Std 0.855308 +eval/Actions Max 0.999973 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5938 +time/logging (s) 0.00373628 +time/sampling batch (s) 0.52958 +time/saving (s) 0.00347027 +time/training (s) 6.72677 +time/epoch (s) 9.85736 +time/total (s) 9780.84 +Epoch -38 +---------------------------------- --------------- +2022-05-10 15:54:02.740427 PDT | [0] Epoch -37 finished +---------------------------------- --------------- +epoch -37 +replay_buffer/size 999033 +trainer/num train calls 964000 +trainer/Policy Loss -19.3566 +trainer/Log Pis Mean 23.3794 +trainer/Log Pis Std 12.803 +trainer/Log Pis Max 65.2156 +trainer/Log Pis Min -6.27683 +trainer/policy/mean Mean -0.0210162 +trainer/policy/mean Std 0.90608 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999985 +trainer/policy/normal/std Mean 2.82195 +trainer/policy/normal/std Std 0.651924 +trainer/policy/normal/std Max 6.61369 +trainer/policy/normal/std Min 0.308199 +trainer/policy/normal/log_std Mean 1.00033 +trainer/policy/normal/log_std Std 0.304368 +trainer/policy/normal/log_std Max 1.88914 +trainer/policy/normal/log_std Min -1.17701 +eval/num steps total 960320 +eval/num paths total 964 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.138157 +eval/Actions Std 0.923412 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42105 +time/logging (s) 0.00371237 +time/sampling batch (s) 0.278879 +time/saving (s) 0.00345112 +time/training (s) 7.7487 +time/epoch (s) 10.4558 +time/total (s) 9791.3 +Epoch -37 +---------------------------------- --------------- +2022-05-10 15:54:14.973804 PDT | [0] Epoch -36 finished +---------------------------------- --------------- +epoch -36 +replay_buffer/size 999033 +trainer/num train calls 965000 +trainer/Policy Loss -19.6995 +trainer/Log Pis Mean 24.2169 +trainer/Log Pis Std 13.0867 +trainer/Log Pis Max 67.2546 +trainer/Log Pis Min -4.91163 +trainer/policy/mean Mean -0.047316 +trainer/policy/mean Std 0.909327 +trainer/policy/mean Max 1 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.79184 +trainer/policy/normal/std Std 0.658716 +trainer/policy/normal/std Max 5.80885 +trainer/policy/normal/std Min 0.296492 +trainer/policy/normal/log_std Mean 0.986929 +trainer/policy/normal/log_std Std 0.317531 +trainer/policy/normal/log_std Max 1.75938 +trainer/policy/normal/log_std Min -1.21573 +eval/num steps total 961320 +eval/num paths total 965 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.106973 +eval/Actions Std 0.920325 +eval/Actions Max 0.999999 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60985 +time/logging (s) 0.00369709 +time/sampling batch (s) 0.781005 +time/saving (s) 0.0035457 +time/training (s) 8.81089 +time/epoch (s) 12.209 +time/total (s) 9803.51 +Epoch -36 +---------------------------------- --------------- +2022-05-10 15:54:24.878650 PDT | [0] Epoch -35 finished +---------------------------------- --------------- +epoch -35 +replay_buffer/size 999033 +trainer/num train calls 966000 +trainer/Policy Loss -19.8253 +trainer/Log Pis Mean 23.572 +trainer/Log Pis Std 12.9451 +trainer/Log Pis Max 63.6371 +trainer/Log Pis Min -12.7874 +trainer/policy/mean Mean -0.0212081 +trainer/policy/mean Std 0.904136 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76437 +trainer/policy/normal/std Std 0.696774 +trainer/policy/normal/std Max 6.67336 +trainer/policy/normal/std Min 0.24705 +trainer/policy/normal/log_std Mean 0.970866 +trainer/policy/normal/log_std Std 0.343016 +trainer/policy/normal/log_std Max 1.89812 +trainer/policy/normal/log_std Min -1.39817 +eval/num steps total 962320 +eval/num paths total 966 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.432214 +eval/Actions Std 0.80828 +eval/Actions Max 0.999989 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.5126 +time/logging (s) 0.00369472 +time/sampling batch (s) 0.529119 +time/saving (s) 0.00344608 +time/training (s) 6.83178 +time/epoch (s) 9.88064 +time/total (s) 9813.39 +Epoch -35 +---------------------------------- --------------- +2022-05-10 15:54:35.002287 PDT | [0] Epoch -34 finished +---------------------------------- --------------- +epoch -34 +replay_buffer/size 999033 +trainer/num train calls 967000 +trainer/Policy Loss -19.313 +trainer/Log Pis Mean 23.6928 +trainer/Log Pis Std 13.7971 +trainer/Log Pis Max 76.9493 +trainer/Log Pis Min -6.70436 +trainer/policy/mean Mean -0.0296183 +trainer/policy/mean Std 0.907052 +trainer/policy/mean Max 0.999983 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.76859 +trainer/policy/normal/std Std 0.685415 +trainer/policy/normal/std Max 7.36195 +trainer/policy/normal/std Min 0.252783 +trainer/policy/normal/log_std Mean 0.970358 +trainer/policy/normal/log_std Std 0.358853 +trainer/policy/normal/log_std Max 1.99633 +trainer/policy/normal/log_std Min -1.37522 +eval/num steps total 963320 +eval/num paths total 967 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0278273 +eval/Actions Std 0.906365 +eval/Actions Max 0.999997 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.51334 +time/logging (s) 0.00370582 +time/sampling batch (s) 0.278186 +time/saving (s) 0.00345438 +time/training (s) 7.30069 +time/epoch (s) 10.0994 +time/total (s) 9823.49 +Epoch -34 +---------------------------------- --------------- +2022-05-10 15:54:44.577819 PDT | [0] Epoch -33 finished +---------------------------------- --------------- +epoch -33 +replay_buffer/size 999033 +trainer/num train calls 968000 +trainer/Policy Loss -20.0734 +trainer/Log Pis Mean 24.9153 +trainer/Log Pis Std 12.6255 +trainer/Log Pis Max 60.237 +trainer/Log Pis Min -9.70404 +trainer/policy/mean Mean -0.0247367 +trainer/policy/mean Std 0.905448 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.777 +trainer/policy/normal/std Std 0.712431 +trainer/policy/normal/std Max 5.87679 +trainer/policy/normal/std Min 0.310223 +trainer/policy/normal/log_std Mean 0.972922 +trainer/policy/normal/log_std Std 0.354338 +trainer/policy/normal/log_std Max 1.77101 +trainer/policy/normal/log_std Min -1.17046 +eval/num steps total 964320 +eval/num paths total 968 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.287257 +eval/Actions Std 0.835001 +eval/Actions Max 0.99997 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.76889 +time/logging (s) 0.00372238 +time/sampling batch (s) 0.281357 +time/saving (s) 0.00345909 +time/training (s) 6.49359 +time/epoch (s) 9.55102 +time/total (s) 9833.05 +Epoch -33 +---------------------------------- --------------- +2022-05-10 15:54:54.423146 PDT | [0] Epoch -32 finished +---------------------------------- --------------- +epoch -32 +replay_buffer/size 999033 +trainer/num train calls 969000 +trainer/Policy Loss -19.0549 +trainer/Log Pis Mean 23.888 +trainer/Log Pis Std 12.3964 +trainer/Log Pis Max 69.7199 +trainer/Log Pis Min -5.37285 +trainer/policy/mean Mean -0.0418376 +trainer/policy/mean Std 0.907714 +trainer/policy/mean Max 0.999998 +trainer/policy/mean Min -0.99999 +trainer/policy/normal/std Mean 2.74428 +trainer/policy/normal/std Std 0.648224 +trainer/policy/normal/std Max 5.22495 +trainer/policy/normal/std Min 0.262441 +trainer/policy/normal/log_std Mean 0.969198 +trainer/policy/normal/log_std Std 0.320519 +trainer/policy/normal/log_std Max 1.65345 +trainer/policy/normal/log_std Min -1.33773 +eval/num steps total 965320 +eval/num paths total 969 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.608083 +eval/Actions Std 0.766298 +eval/Actions Max 0.999994 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.50234 +time/logging (s) 0.00484897 +time/sampling batch (s) 0.280162 +time/saving (s) 0.00418515 +time/training (s) 7.03048 +time/epoch (s) 9.82202 +time/total (s) 9842.87 +Epoch -32 +---------------------------------- --------------- +2022-05-10 15:55:04.303876 PDT | [0] Epoch -31 finished +---------------------------------- --------------- +epoch -31 +replay_buffer/size 999033 +trainer/num train calls 970000 +trainer/Policy Loss -19.1999 +trainer/Log Pis Mean 25.1608 +trainer/Log Pis Std 13.1861 +trainer/Log Pis Max 77.3589 +trainer/Log Pis Min -10.1951 +trainer/policy/mean Mean -0.0197069 +trainer/policy/mean Std 0.906598 +trainer/policy/mean Max 0.999981 +trainer/policy/mean Min -0.999977 +trainer/policy/normal/std Mean 2.77047 +trainer/policy/normal/std Std 0.658881 +trainer/policy/normal/std Max 5.83276 +trainer/policy/normal/std Min 0.283223 +trainer/policy/normal/log_std Mean 0.978063 +trainer/policy/normal/log_std Std 0.323294 +trainer/policy/normal/log_std Max 1.76349 +trainer/policy/normal/log_std Min -1.26152 +eval/num steps total 966320 +eval/num paths total 970 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0620693 +eval/Actions Std 0.917768 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53164 +time/logging (s) 0.00374803 +time/sampling batch (s) 0.531672 +time/saving (s) 0.0036091 +time/training (s) 6.78322 +time/epoch (s) 9.85389 +time/total (s) 9852.73 +Epoch -31 +---------------------------------- --------------- +2022-05-10 15:55:13.841424 PDT | [0] Epoch -30 finished +---------------------------------- --------------- +epoch -30 +replay_buffer/size 999033 +trainer/num train calls 971000 +trainer/Policy Loss -21.2974 +trainer/Log Pis Mean 24.5578 +trainer/Log Pis Std 13.0842 +trainer/Log Pis Max 67.7889 +trainer/Log Pis Min -7.29961 +trainer/policy/mean Mean -0.0310654 +trainer/policy/mean Std 0.912148 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.80536 +trainer/policy/normal/std Std 0.662319 +trainer/policy/normal/std Max 5.2812 +trainer/policy/normal/std Min 0.316638 +trainer/policy/normal/log_std Mean 0.990456 +trainer/policy/normal/log_std Std 0.325611 +trainer/policy/normal/log_std Max 1.66415 +trainer/policy/normal/log_std Min -1.15 +eval/num steps total 967320 +eval/num paths total 971 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0718952 +eval/Actions Std 0.916828 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.49515 +time/logging (s) 0.00377799 +time/sampling batch (s) 0.280333 +time/saving (s) 0.00347091 +time/training (s) 6.73037 +time/epoch (s) 9.5131 +time/total (s) 9862.25 +Epoch -30 +---------------------------------- --------------- +2022-05-10 15:55:23.696977 PDT | [0] Epoch -29 finished +---------------------------------- --------------- +epoch -29 +replay_buffer/size 999033 +trainer/num train calls 972000 +trainer/Policy Loss -20.1212 +trainer/Log Pis Mean 24.2068 +trainer/Log Pis Std 13.4361 +trainer/Log Pis Max 66.0831 +trainer/Log Pis Min -10.9841 +trainer/policy/mean Mean -0.0410739 +trainer/policy/mean Std 0.906956 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.77728 +trainer/policy/normal/std Std 0.666632 +trainer/policy/normal/std Max 6.32919 +trainer/policy/normal/std Min 0.292909 +trainer/policy/normal/log_std Mean 0.978554 +trainer/policy/normal/log_std Std 0.333817 +trainer/policy/normal/log_std Max 1.84517 +trainer/policy/normal/log_std Min -1.22789 +eval/num steps total 968320 +eval/num paths total 972 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.398147 +eval/Actions Std 0.835717 +eval/Actions Max 0.999997 +eval/Actions Min -0.999992 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.89276 +time/logging (s) 0.00390848 +time/sampling batch (s) 0.53434 +time/saving (s) 0.0035518 +time/training (s) 6.39641 +time/epoch (s) 9.83097 +time/total (s) 9872.08 +Epoch -29 +---------------------------------- --------------- +2022-05-10 15:55:34.838432 PDT | [0] Epoch -28 finished +---------------------------------- --------------- +epoch -28 +replay_buffer/size 999033 +trainer/num train calls 973000 +trainer/Policy Loss -18.3346 +trainer/Log Pis Mean 24.3161 +trainer/Log Pis Std 13.2818 +trainer/Log Pis Max 74.6835 +trainer/Log Pis Min -6.69528 +trainer/policy/mean Mean -0.0547128 +trainer/policy/mean Std 0.910788 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77581 +trainer/policy/normal/std Std 0.638561 +trainer/policy/normal/std Max 5.32876 +trainer/policy/normal/std Min 0.282704 +trainer/policy/normal/log_std Mean 0.983421 +trainer/policy/normal/log_std Std 0.307732 +trainer/policy/normal/log_std Max 1.67312 +trainer/policy/normal/log_std Min -1.26336 +eval/num steps total 969320 +eval/num paths total 973 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.360428 +eval/Actions Std 0.853 +eval/Actions Max 0.999992 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64488 +time/logging (s) 0.00366672 +time/sampling batch (s) 0.290649 +time/saving (s) 0.00348641 +time/training (s) 8.17329 +time/epoch (s) 11.116 +time/total (s) 9883.2 +Epoch -28 +---------------------------------- --------------- +2022-05-10 15:55:44.360953 PDT | [0] Epoch -27 finished +---------------------------------- --------------- +epoch -27 +replay_buffer/size 999033 +trainer/num train calls 974000 +trainer/Policy Loss -17.9513 +trainer/Log Pis Mean 23.2321 +trainer/Log Pis Std 12.9228 +trainer/Log Pis Max 66.745 +trainer/Log Pis Min -7.83121 +trainer/policy/mean Mean -0.0201702 +trainer/policy/mean Std 0.902145 +trainer/policy/mean Max 0.99998 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.76469 +trainer/policy/normal/std Std 0.694534 +trainer/policy/normal/std Max 5.97397 +trainer/policy/normal/std Min 0.216431 +trainer/policy/normal/log_std Mean 0.968767 +trainer/policy/normal/log_std Std 0.356589 +trainer/policy/normal/log_std Max 1.78741 +trainer/policy/normal/log_std Min -1.53048 +eval/num steps total 970320 +eval/num paths total 974 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.128055 +eval/Actions Std 0.920755 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58113 +time/logging (s) 0.0038235 +time/sampling batch (s) 0.283412 +time/saving (s) 0.00342355 +time/training (s) 6.62648 +time/epoch (s) 9.49827 +time/total (s) 9892.7 +Epoch -27 +---------------------------------- --------------- +2022-05-10 15:55:54.048648 PDT | [0] Epoch -26 finished +---------------------------------- --------------- +epoch -26 +replay_buffer/size 999033 +trainer/num train calls 975000 +trainer/Policy Loss -19.8528 +trainer/Log Pis Mean 24.379 +trainer/Log Pis Std 13.0332 +trainer/Log Pis Max 60.6782 +trainer/Log Pis Min -7.79682 +trainer/policy/mean Mean -0.0371145 +trainer/policy/mean Std 0.905454 +trainer/policy/mean Max 0.999976 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.80728 +trainer/policy/normal/std Std 0.694436 +trainer/policy/normal/std Max 5.66043 +trainer/policy/normal/std Min 0.265278 +trainer/policy/normal/log_std Mean 0.986252 +trainer/policy/normal/log_std Std 0.347186 +trainer/policy/normal/log_std Max 1.7335 +trainer/policy/normal/log_std Min -1.32698 +eval/num steps total 971320 +eval/num paths total 975 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0611976 +eval/Actions Std 0.919526 +eval/Actions Max 1 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42159 +time/logging (s) 0.00372361 +time/sampling batch (s) 0.284217 +time/saving (s) 0.0034156 +time/training (s) 6.95028 +time/epoch (s) 9.66322 +time/total (s) 9902.37 +Epoch -26 +---------------------------------- --------------- +2022-05-10 15:56:04.002169 PDT | [0] Epoch -25 finished +---------------------------------- --------------- +epoch -25 +replay_buffer/size 999033 +trainer/num train calls 976000 +trainer/Policy Loss -19.7925 +trainer/Log Pis Mean 23.765 +trainer/Log Pis Std 13.0627 +trainer/Log Pis Max 71.1385 +trainer/Log Pis Min -8.05593 +trainer/policy/mean Mean -0.024643 +trainer/policy/mean Std 0.904195 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999987 +trainer/policy/normal/std Mean 2.76681 +trainer/policy/normal/std Std 0.674602 +trainer/policy/normal/std Max 5.71223 +trainer/policy/normal/std Min 0.278153 +trainer/policy/normal/log_std Mean 0.974498 +trainer/policy/normal/log_std Std 0.332628 +trainer/policy/normal/log_std Max 1.74261 +trainer/policy/normal/log_std Min -1.27958 +eval/num steps total 972320 +eval/num paths total 976 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0795927 +eval/Actions Std 0.898735 +eval/Actions Max 0.999999 +eval/Actions Min -1 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62079 +time/logging (s) 0.00391394 +time/sampling batch (s) 0.279732 +time/saving (s) 0.00411909 +time/training (s) 7.02095 +time/epoch (s) 9.9295 +time/total (s) 9912.3 +Epoch -25 +---------------------------------- --------------- +2022-05-10 15:56:13.964543 PDT | [0] Epoch -24 finished +---------------------------------- --------------- +epoch -24 +replay_buffer/size 999033 +trainer/num train calls 977000 +trainer/Policy Loss -19.9295 +trainer/Log Pis Mean 25.7824 +trainer/Log Pis Std 13.9412 +trainer/Log Pis Max 68.8767 +trainer/Log Pis Min -4.89342 +trainer/policy/mean Mean -0.0501894 +trainer/policy/mean Std 0.911552 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999997 +trainer/policy/normal/std Mean 2.79139 +trainer/policy/normal/std Std 0.689866 +trainer/policy/normal/std Max 5.68406 +trainer/policy/normal/std Min 0.248369 +trainer/policy/normal/log_std Mean 0.97996 +trainer/policy/normal/log_std Std 0.351133 +trainer/policy/normal/log_std Max 1.73766 +trainer/policy/normal/log_std Min -1.39284 +eval/num steps total 973320 +eval/num paths total 977 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0652009 +eval/Actions Std 0.880761 +eval/Actions Max 0.999987 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.63481 +time/logging (s) 0.00396927 +time/sampling batch (s) 0.284733 +time/saving (s) 0.00347886 +time/training (s) 7.0104 +time/epoch (s) 9.93739 +time/total (s) 9922.24 +Epoch -24 +---------------------------------- --------------- +2022-05-10 15:56:25.184029 PDT | [0] Epoch -23 finished +---------------------------------- --------------- +epoch -23 +replay_buffer/size 999033 +trainer/num train calls 978000 +trainer/Policy Loss -18.4415 +trainer/Log Pis Mean 23.6651 +trainer/Log Pis Std 13.0351 +trainer/Log Pis Max 64.4202 +trainer/Log Pis Min -10.1059 +trainer/policy/mean Mean -0.0407621 +trainer/policy/mean Std 0.901682 +trainer/policy/mean Max 0.999986 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79427 +trainer/policy/normal/std Std 0.676594 +trainer/policy/normal/std Max 6.56595 +trainer/policy/normal/std Min 0.361687 +trainer/policy/normal/log_std Mean 0.986017 +trainer/policy/normal/log_std Std 0.323828 +trainer/policy/normal/log_std Max 1.8819 +trainer/policy/normal/log_std Min -1.01698 +eval/num steps total 974320 +eval/num paths total 978 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.217309 +eval/Actions Std 0.763701 +eval/Actions Max 0.999993 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.67954 +time/logging (s) 0.00370195 +time/sampling batch (s) 0.286942 +time/saving (s) 0.003371 +time/training (s) 8.22052 +time/epoch (s) 11.1941 +time/total (s) 9933.43 +Epoch -23 +---------------------------------- --------------- +2022-05-10 15:56:34.722362 PDT | [0] Epoch -22 finished +---------------------------------- --------------- +epoch -22 +replay_buffer/size 999033 +trainer/num train calls 979000 +trainer/Policy Loss -19.8899 +trainer/Log Pis Mean 24.6461 +trainer/Log Pis Std 12.8933 +trainer/Log Pis Max 63.5939 +trainer/Log Pis Min -5.99843 +trainer/policy/mean Mean -0.0335324 +trainer/policy/mean Std 0.907829 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.77096 +trainer/policy/normal/std Std 0.664553 +trainer/policy/normal/std Max 5.72809 +trainer/policy/normal/std Min 0.238502 +trainer/policy/normal/log_std Mean 0.975135 +trainer/policy/normal/log_std Std 0.342829 +trainer/policy/normal/log_std Max 1.74538 +trainer/policy/normal/log_std Min -1.43338 +eval/num steps total 975320 +eval/num paths total 979 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.381523 +eval/Actions Std 0.835038 +eval/Actions Max 0.999997 +eval/Actions Min -0.999995 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57692 +time/logging (s) 0.00368076 +time/sampling batch (s) 0.281895 +time/saving (s) 0.00341338 +time/training (s) 6.64786 +time/epoch (s) 9.51377 +time/total (s) 9942.95 +Epoch -22 +---------------------------------- --------------- +2022-05-10 15:56:44.583095 PDT | [0] Epoch -21 finished +---------------------------------- --------------- +epoch -21 +replay_buffer/size 999033 +trainer/num train calls 980000 +trainer/Policy Loss -19.9795 +trainer/Log Pis Mean 25.2708 +trainer/Log Pis Std 13.0848 +trainer/Log Pis Max 73.8751 +trainer/Log Pis Min -11.4549 +trainer/policy/mean Mean -0.0530707 +trainer/policy/mean Std 0.90884 +trainer/policy/mean Max 0.999984 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.78059 +trainer/policy/normal/std Std 0.666721 +trainer/policy/normal/std Max 6.29039 +trainer/policy/normal/std Min 0.299451 +trainer/policy/normal/log_std Mean 0.981405 +trainer/policy/normal/log_std Std 0.32483 +trainer/policy/normal/log_std Max 1.83902 +trainer/policy/normal/log_std Min -1.2058 +eval/num steps total 976320 +eval/num paths total 980 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.132992 +eval/Actions Std 0.882916 +eval/Actions Max 0.999999 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.70277 +time/logging (s) 0.00392166 +time/sampling batch (s) 0.284483 +time/saving (s) 0.003456 +time/training (s) 6.84155 +time/epoch (s) 9.83619 +time/total (s) 9952.79 +Epoch -21 +---------------------------------- --------------- +2022-05-10 15:56:56.251971 PDT | [0] Epoch -20 finished +---------------------------------- --------------- +epoch -20 +replay_buffer/size 999033 +trainer/num train calls 981000 +trainer/Policy Loss -20.8033 +trainer/Log Pis Mean 23.4143 +trainer/Log Pis Std 12.8745 +trainer/Log Pis Max 69.3157 +trainer/Log Pis Min -7.72218 +trainer/policy/mean Mean -0.0317242 +trainer/policy/mean Std 0.905641 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.76902 +trainer/policy/normal/std Std 0.683012 +trainer/policy/normal/std Max 6.06671 +trainer/policy/normal/std Min 0.238778 +trainer/policy/normal/log_std Mean 0.97065 +trainer/policy/normal/log_std Std 0.359547 +trainer/policy/normal/log_std Max 1.80282 +trainer/policy/normal/log_std Min -1.43222 +eval/num steps total 977320 +eval/num paths total 981 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.163879 +eval/Actions Std 0.936749 +eval/Actions Max 0.999997 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.62948 +time/logging (s) 0.00434603 +time/sampling batch (s) 0.28597 +time/saving (s) 0.00436239 +time/training (s) 8.71873 +time/epoch (s) 11.6429 +time/total (s) 9964.44 +Epoch -20 +---------------------------------- --------------- +2022-05-10 15:57:06.811850 PDT | [0] Epoch -19 finished +---------------------------------- --------------- +epoch -19 +replay_buffer/size 999033 +trainer/num train calls 982000 +trainer/Policy Loss -18.7392 +trainer/Log Pis Mean 24.5504 +trainer/Log Pis Std 13.3487 +trainer/Log Pis Max 67.4914 +trainer/Log Pis Min -6.03882 +trainer/policy/mean Mean -0.0298839 +trainer/policy/mean Std 0.904786 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.81388 +trainer/policy/normal/std Std 0.684088 +trainer/policy/normal/std Max 6.18759 +trainer/policy/normal/std Min 0.274495 +trainer/policy/normal/log_std Mean 0.989989 +trainer/policy/normal/log_std Std 0.343318 +trainer/policy/normal/log_std Max 1.82255 +trainer/policy/normal/log_std Min -1.29282 +eval/num steps total 978320 +eval/num paths total 982 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.209743 +eval/Actions Std 0.962974 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58061 +time/logging (s) 0.00380363 +time/sampling batch (s) 0.533442 +time/saving (s) 0.00341727 +time/training (s) 7.41286 +time/epoch (s) 10.5341 +time/total (s) 9974.97 +Epoch -19 +---------------------------------- --------------- +2022-05-10 15:57:17.451956 PDT | [0] Epoch -18 finished +---------------------------------- --------------- +epoch -18 +replay_buffer/size 999033 +trainer/num train calls 983000 +trainer/Policy Loss -20.3357 +trainer/Log Pis Mean 26.0172 +trainer/Log Pis Std 13.2315 +trainer/Log Pis Max 69.4691 +trainer/Log Pis Min -6.81753 +trainer/policy/mean Mean -0.0492632 +trainer/policy/mean Std 0.908015 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.85338 +trainer/policy/normal/std Std 0.693241 +trainer/policy/normal/std Max 7.38906 +trainer/policy/normal/std Min 0.303626 +trainer/policy/normal/log_std Mean 1.00747 +trainer/policy/normal/log_std Std 0.319568 +trainer/policy/normal/log_std Max 2 +trainer/policy/normal/log_std Min -1.19196 +eval/num steps total 979320 +eval/num paths total 983 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.116739 +eval/Actions Std 0.831866 +eval/Actions Max 0.999998 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.45758 +time/logging (s) 0.00381827 +time/sampling batch (s) 0.283109 +time/saving (s) 0.00350069 +time/training (s) 7.86749 +time/epoch (s) 10.6155 +time/total (s) 9985.59 +Epoch -18 +---------------------------------- --------------- +2022-05-10 15:57:26.850848 PDT | [0] Epoch -17 finished +---------------------------------- --------------- +epoch -17 +replay_buffer/size 999033 +trainer/num train calls 984000 +trainer/Policy Loss -19.5269 +trainer/Log Pis Mean 23.7319 +trainer/Log Pis Std 13.0226 +trainer/Log Pis Max 74.7751 +trainer/Log Pis Min -8.24713 +trainer/policy/mean Mean -0.0476295 +trainer/policy/mean Std 0.904664 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.80675 +trainer/policy/normal/std Std 0.665649 +trainer/policy/normal/std Max 5.27621 +trainer/policy/normal/std Min 0.241313 +trainer/policy/normal/log_std Mean 0.990477 +trainer/policy/normal/log_std Std 0.327593 +trainer/policy/normal/log_std Max 1.66321 +trainer/policy/normal/log_std Min -1.42166 +eval/num steps total 980320 +eval/num paths total 984 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.244923 +eval/Actions Std 0.961884 +eval/Actions Max 0.999973 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.40768 +time/logging (s) 0.00407056 +time/sampling batch (s) 0.283373 +time/saving (s) 0.0040383 +time/training (s) 6.67527 +time/epoch (s) 9.37444 +time/total (s) 9994.97 +Epoch -17 +---------------------------------- --------------- +2022-05-10 15:57:36.814490 PDT | [0] Epoch -16 finished +---------------------------------- --------------- +epoch -16 +replay_buffer/size 999033 +trainer/num train calls 985000 +trainer/Policy Loss -19.3277 +trainer/Log Pis Mean 24.8737 +trainer/Log Pis Std 13.4044 +trainer/Log Pis Max 71.4516 +trainer/Log Pis Min -10.6026 +trainer/policy/mean Mean -0.0544922 +trainer/policy/mean Std 0.905395 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.86197 +trainer/policy/normal/std Std 0.693906 +trainer/policy/normal/std Max 6.25109 +trainer/policy/normal/std Min 0.260612 +trainer/policy/normal/log_std Mean 1.00716 +trainer/policy/normal/log_std Std 0.342203 +trainer/policy/normal/log_std Max 1.83276 +trainer/policy/normal/log_std Min -1.34472 +eval/num steps total 981320 +eval/num paths total 985 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0218582 +eval/Actions Std 0.898111 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61682 +time/logging (s) 0.00384413 +time/sampling batch (s) 0.533344 +time/saving (s) 0.00370175 +time/training (s) 6.78053 +time/epoch (s) 9.93824 +time/total (s) 10004.9 +Epoch -16 +---------------------------------- --------------- +2022-05-10 15:57:47.292613 PDT | [0] Epoch -15 finished +---------------------------------- --------------- +epoch -15 +replay_buffer/size 999033 +trainer/num train calls 986000 +trainer/Policy Loss -18.9419 +trainer/Log Pis Mean 23.9117 +trainer/Log Pis Std 13.1367 +trainer/Log Pis Max 64.5321 +trainer/Log Pis Min -5.82417 +trainer/policy/mean Mean -0.011671 +trainer/policy/mean Std 0.90754 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999984 +trainer/policy/normal/std Mean 2.75958 +trainer/policy/normal/std Std 0.677223 +trainer/policy/normal/std Max 6.34203 +trainer/policy/normal/std Min 0.268307 +trainer/policy/normal/log_std Mean 0.970232 +trainer/policy/normal/log_std Std 0.341608 +trainer/policy/normal/log_std Max 1.8472 +trainer/policy/normal/log_std Min -1.31563 +eval/num steps total 982320 +eval/num paths total 986 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.138732 +eval/Actions Std 0.827176 +eval/Actions Max 0.999998 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.58225 +time/logging (s) 0.00368967 +time/sampling batch (s) 0.286707 +time/saving (s) 0.00359952 +time/training (s) 7.57668 +time/epoch (s) 10.4529 +time/total (s) 10015.4 +Epoch -15 +---------------------------------- --------------- +2022-05-10 15:57:56.632574 PDT | [0] Epoch -14 finished +---------------------------------- --------------- +epoch -14 +replay_buffer/size 999033 +trainer/num train calls 987000 +trainer/Policy Loss -20.351 +trainer/Log Pis Mean 25.4158 +trainer/Log Pis Std 13.8098 +trainer/Log Pis Max 78.3875 +trainer/Log Pis Min -8.42511 +trainer/policy/mean Mean -0.0459769 +trainer/policy/mean Std 0.908449 +trainer/policy/mean Max 0.999989 +trainer/policy/mean Min -0.999995 +trainer/policy/normal/std Mean 2.82754 +trainer/policy/normal/std Std 0.682598 +trainer/policy/normal/std Max 5.78853 +trainer/policy/normal/std Min 0.291281 +trainer/policy/normal/log_std Mean 0.995921 +trainer/policy/normal/log_std Std 0.336493 +trainer/policy/normal/log_std Max 1.75588 +trainer/policy/normal/log_std Min -1.23347 +eval/num steps total 983320 +eval/num paths total 987 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.269322 +eval/Actions Std 0.876807 +eval/Actions Max 0.99999 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.6419 +time/logging (s) 0.00374783 +time/sampling batch (s) 0.278982 +time/saving (s) 0.00343907 +time/training (s) 6.38766 +time/epoch (s) 9.31573 +time/total (s) 10024.7 +Epoch -14 +---------------------------------- --------------- +2022-05-10 15:58:07.047907 PDT | [0] Epoch -13 finished +---------------------------------- --------------- +epoch -13 +replay_buffer/size 999033 +trainer/num train calls 988000 +trainer/Policy Loss -19.5067 +trainer/Log Pis Mean 23.9455 +trainer/Log Pis Std 13.7386 +trainer/Log Pis Max 74.4509 +trainer/Log Pis Min -12.7017 +trainer/policy/mean Mean -0.0314157 +trainer/policy/mean Std 0.905035 +trainer/policy/mean Max 0.999987 +trainer/policy/mean Min -0.999986 +trainer/policy/normal/std Mean 2.77249 +trainer/policy/normal/std Std 0.653367 +trainer/policy/normal/std Max 5.39209 +trainer/policy/normal/std Min 0.341486 +trainer/policy/normal/log_std Mean 0.97916 +trainer/policy/normal/log_std Std 0.322935 +trainer/policy/normal/log_std Max 1.68493 +trainer/policy/normal/log_std Min -1.07445 +eval/num steps total 984320 +eval/num paths total 988 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.107987 +eval/Actions Std 0.894058 +eval/Actions Max 0.999996 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57914 +time/logging (s) 0.00406406 +time/sampling batch (s) 0.280107 +time/saving (s) 0.00414685 +time/training (s) 7.5238 +time/epoch (s) 10.3913 +time/total (s) 10035.1 +Epoch -13 +---------------------------------- --------------- +2022-05-10 15:58:16.719501 PDT | [0] Epoch -12 finished +---------------------------------- --------------- +epoch -12 +replay_buffer/size 999033 +trainer/num train calls 989000 +trainer/Policy Loss -19.6412 +trainer/Log Pis Mean 24.4456 +trainer/Log Pis Std 13.5034 +trainer/Log Pis Max 74.5043 +trainer/Log Pis Min -9.44313 +trainer/policy/mean Mean -0.0326934 +trainer/policy/mean Std 0.910906 +trainer/policy/mean Max 0.999999 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.87288 +trainer/policy/normal/std Std 0.682662 +trainer/policy/normal/std Max 6.32151 +trainer/policy/normal/std Min 0.279378 +trainer/policy/normal/log_std Mean 1.01431 +trainer/policy/normal/log_std Std 0.324844 +trainer/policy/normal/log_std Max 1.84396 +trainer/policy/normal/log_std Min -1.27519 +eval/num steps total 985320 +eval/num paths total 989 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.095798 +eval/Actions Std 0.917536 +eval/Actions Max 0.999993 +eval/Actions Min -0.999996 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57818 +time/logging (s) 0.00380007 +time/sampling batch (s) 0.285848 +time/saving (s) 0.00347644 +time/training (s) 6.77482 +time/epoch (s) 9.64612 +time/total (s) 10044.7 +Epoch -12 +---------------------------------- --------------- +2022-05-10 15:58:26.383466 PDT | [0] Epoch -11 finished +---------------------------------- --------------- +epoch -11 +replay_buffer/size 999033 +trainer/num train calls 990000 +trainer/Policy Loss -19.8591 +trainer/Log Pis Mean 24.0506 +trainer/Log Pis Std 13.5073 +trainer/Log Pis Max 66.9719 +trainer/Log Pis Min -8.71334 +trainer/policy/mean Mean -0.023403 +trainer/policy/mean Std 0.909974 +trainer/policy/mean Max 0.999996 +trainer/policy/mean Min -0.999994 +trainer/policy/normal/std Mean 2.76032 +trainer/policy/normal/std Std 0.673678 +trainer/policy/normal/std Max 6.0281 +trainer/policy/normal/std Min 0.293858 +trainer/policy/normal/log_std Mean 0.969924 +trainer/policy/normal/log_std Std 0.346094 +trainer/policy/normal/log_std Max 1.79643 +trainer/policy/normal/log_std Min -1.22466 +eval/num steps total 986320 +eval/num paths total 990 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0527179 +eval/Actions Std 0.873435 +eval/Actions Max 0.999982 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.74088 +time/logging (s) 0.0043908 +time/sampling batch (s) 0.28284 +time/saving (s) 0.00414895 +time/training (s) 6.60761 +time/epoch (s) 9.63987 +time/total (s) 10054.4 +Epoch -11 +---------------------------------- --------------- +2022-05-10 15:58:36.072511 PDT | [0] Epoch -10 finished +---------------------------------- --------------- +epoch -10 +replay_buffer/size 999033 +trainer/num train calls 991000 +trainer/Policy Loss -20.1785 +trainer/Log Pis Mean 23.9687 +trainer/Log Pis Std 13.2176 +trainer/Log Pis Max 68.2789 +trainer/Log Pis Min -6.50378 +trainer/policy/mean Mean -0.0284267 +trainer/policy/mean Std 0.902206 +trainer/policy/mean Max 0.999997 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.76922 +trainer/policy/normal/std Std 0.700415 +trainer/policy/normal/std Max 7.08991 +trainer/policy/normal/std Min 0.248668 +trainer/policy/normal/log_std Mean 0.970967 +trainer/policy/normal/log_std Std 0.353786 +trainer/policy/normal/log_std Max 1.95867 +trainer/policy/normal/log_std Min -1.39164 +eval/num steps total 987320 +eval/num paths total 991 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.0532137 +eval/Actions Std 0.911359 +eval/Actions Max 1 +eval/Actions Min -0.999997 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57547 +time/logging (s) 0.00377532 +time/sampling batch (s) 0.285459 +time/saving (s) 0.00370293 +time/training (s) 6.79422 +time/epoch (s) 9.66263 +time/total (s) 10064 +Epoch -10 +---------------------------------- --------------- +2022-05-10 15:58:45.409137 PDT | [0] Epoch -9 finished +---------------------------------- --------------- +epoch -9 +replay_buffer/size 999033 +trainer/num train calls 992000 +trainer/Policy Loss -18.7127 +trainer/Log Pis Mean 24.9144 +trainer/Log Pis Std 14.2616 +trainer/Log Pis Max 72.5061 +trainer/Log Pis Min -7.96129 +trainer/policy/mean Mean -0.0383816 +trainer/policy/mean Std 0.902196 +trainer/policy/mean Max 0.999994 +trainer/policy/mean Min -0.999996 +trainer/policy/normal/std Mean 2.81289 +trainer/policy/normal/std Std 0.690042 +trainer/policy/normal/std Max 6.20546 +trainer/policy/normal/std Min 0.32166 +trainer/policy/normal/log_std Mean 0.989822 +trainer/policy/normal/log_std Std 0.339028 +trainer/policy/normal/log_std Max 1.82543 +trainer/policy/normal/log_std Min -1.13426 +eval/num steps total 988320 +eval/num paths total 992 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.366715 +eval/Actions Std 0.851831 +eval/Actions Max 0.999993 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.61448 +time/logging (s) 0.00390458 +time/sampling batch (s) 0.533911 +time/saving (s) 0.00362872 +time/training (s) 6.15602 +time/epoch (s) 9.31195 +time/total (s) 10073.4 +Epoch -9 +---------------------------------- --------------- +2022-05-10 15:58:55.225680 PDT | [0] Epoch -8 finished +---------------------------------- --------------- +epoch -8 +replay_buffer/size 999033 +trainer/num train calls 993000 +trainer/Policy Loss -20.7449 +trainer/Log Pis Mean 24.5881 +trainer/Log Pis Std 13.594 +trainer/Log Pis Max 59.5212 +trainer/Log Pis Min -8.93426 +trainer/policy/mean Mean -0.0472175 +trainer/policy/mean Std 0.899976 +trainer/policy/mean Max 0.999974 +trainer/policy/mean Min -0.999993 +trainer/policy/normal/std Mean 2.74314 +trainer/policy/normal/std Std 0.682081 +trainer/policy/normal/std Max 5.87268 +trainer/policy/normal/std Min 0.238945 +trainer/policy/normal/log_std Mean 0.962553 +trainer/policy/normal/log_std Std 0.349098 +trainer/policy/normal/log_std Max 1.77031 +trainer/policy/normal/log_std Min -1.43152 +eval/num steps total 989320 +eval/num paths total 993 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0912361 +eval/Actions Std 0.922354 +eval/Actions Max 0.999997 +eval/Actions Min -0.999998 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57738 +time/logging (s) 0.0042702 +time/sampling batch (s) 0.535852 +time/saving (s) 0.00448511 +time/training (s) 6.66989 +time/epoch (s) 9.79187 +time/total (s) 10083.1 +Epoch -8 +---------------------------------- --------------- +2022-05-10 15:59:05.668105 PDT | [0] Epoch -7 finished +---------------------------------- --------------- +epoch -7 +replay_buffer/size 999033 +trainer/num train calls 994000 +trainer/Policy Loss -20.6377 +trainer/Log Pis Mean 24.6459 +trainer/Log Pis Std 13.6564 +trainer/Log Pis Max 70.9376 +trainer/Log Pis Min -11.5675 +trainer/policy/mean Mean -0.0295939 +trainer/policy/mean Std 0.903165 +trainer/policy/mean Max 0.999991 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.88544 +trainer/policy/normal/std Std 0.708009 +trainer/policy/normal/std Max 6.33029 +trainer/policy/normal/std Min 0.245964 +trainer/policy/normal/log_std Mean 1.01454 +trainer/policy/normal/log_std Std 0.344065 +trainer/policy/normal/log_std Max 1.84535 +trainer/policy/normal/log_std Min -1.40257 +eval/num steps total 990320 +eval/num paths total 994 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean 0.191692 +eval/Actions Std 0.925525 +eval/Actions Max 0.999998 +eval/Actions Min -0.999991 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.60161 +time/logging (s) 0.00416548 +time/sampling batch (s) 0.531803 +time/saving (s) 0.00441799 +time/training (s) 7.27487 +time/epoch (s) 10.4169 +time/total (s) 10093.6 +Epoch -7 +---------------------------------- --------------- +2022-05-10 15:59:15.497645 PDT | [0] Epoch -6 finished +---------------------------------- --------------- +epoch -6 +replay_buffer/size 999033 +trainer/num train calls 995000 +trainer/Policy Loss -19.8112 +trainer/Log Pis Mean 23.8012 +trainer/Log Pis Std 13.878 +trainer/Log Pis Max 77.5264 +trainer/Log Pis Min -11.8072 +trainer/policy/mean Mean -0.0368929 +trainer/policy/mean Std 0.907755 +trainer/policy/mean Max 0.999993 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.70819 +trainer/policy/normal/std Std 0.677036 +trainer/policy/normal/std Max 5.93393 +trainer/policy/normal/std Min 0.238889 +trainer/policy/normal/log_std Mean 0.948493 +trainer/policy/normal/log_std Std 0.354291 +trainer/policy/normal/log_std Max 1.78069 +trainer/policy/normal/log_std Min -1.43176 +eval/num steps total 991320 +eval/num paths total 995 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0725193 +eval/Actions Std 0.855685 +eval/Actions Max 0.999989 +eval/Actions Min -0.999994 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64245 +time/logging (s) 0.0038539 +time/sampling batch (s) 0.282764 +time/saving (s) 0.00360887 +time/training (s) 6.87109 +time/epoch (s) 9.80377 +time/total (s) 10103.4 +Epoch -6 +---------------------------------- --------------- +2022-05-10 15:59:25.409781 PDT | [0] Epoch -5 finished +---------------------------------- --------------- +epoch -5 +replay_buffer/size 999033 +trainer/num train calls 996000 +trainer/Policy Loss -20.8349 +trainer/Log Pis Mean 25.538 +trainer/Log Pis Std 13.644 +trainer/Log Pis Max 70.3483 +trainer/Log Pis Min -6.49401 +trainer/policy/mean Mean -0.00911865 +trainer/policy/mean Std 0.911048 +trainer/policy/mean Max 0.999988 +trainer/policy/mean Min -0.999988 +trainer/policy/normal/std Mean 2.7973 +trainer/policy/normal/std Std 0.697852 +trainer/policy/normal/std Max 6.09294 +trainer/policy/normal/std Min 0.221294 +trainer/policy/normal/log_std Mean 0.979204 +trainer/policy/normal/log_std Std 0.365667 +trainer/policy/normal/log_std Max 1.80713 +trainer/policy/normal/log_std Min -1.50826 +eval/num steps total 992320 +eval/num paths total 996 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.468831 +eval/Actions Std 0.733863 +eval/Actions Max 0.999982 +eval/Actions Min -0.999971 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.57043 +time/logging (s) 0.00428136 +time/sampling batch (s) 0.532901 +time/saving (s) 0.00429049 +time/training (s) 6.77564 +time/epoch (s) 9.88755 +time/total (s) 10113.3 +Epoch -5 +---------------------------------- --------------- +2022-05-10 15:59:35.998473 PDT | [0] Epoch -4 finished +---------------------------------- --------------- +epoch -4 +replay_buffer/size 999033 +trainer/num train calls 997000 +trainer/Policy Loss -18.237 +trainer/Log Pis Mean 23.8548 +trainer/Log Pis Std 13.2103 +trainer/Log Pis Max 67.2235 +trainer/Log Pis Min -7.2202 +trainer/policy/mean Mean -0.0354315 +trainer/policy/mean Std 0.905023 +trainer/policy/mean Max 0.999982 +trainer/policy/mean Min -1 +trainer/policy/normal/std Mean 2.81144 +trainer/policy/normal/std Std 0.683881 +trainer/policy/normal/std Max 5.95958 +trainer/policy/normal/std Min 0.249696 +trainer/policy/normal/log_std Mean 0.989178 +trainer/policy/normal/log_std Std 0.341625 +trainer/policy/normal/log_std Max 1.785 +trainer/policy/normal/log_std Min -1.38751 +eval/num steps total 993320 +eval/num paths total 997 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.118677 +eval/Actions Std 0.920811 +eval/Actions Max 0.999986 +eval/Actions Min -0.999979 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53103 +time/logging (s) 0.00369406 +time/sampling batch (s) 0.282501 +time/saving (s) 0.00359197 +time/training (s) 7.74225 +time/epoch (s) 10.5631 +time/total (s) 10123.8 +Epoch -4 +---------------------------------- --------------- +2022-05-10 15:59:45.493005 PDT | [0] Epoch -3 finished +---------------------------------- --------------- +epoch -3 +replay_buffer/size 999033 +trainer/num train calls 998000 +trainer/Policy Loss -19.2537 +trainer/Log Pis Mean 25.4807 +trainer/Log Pis Std 13.199 +trainer/Log Pis Max 66.556 +trainer/Log Pis Min -10.7941 +trainer/policy/mean Mean -0.0133095 +trainer/policy/mean Std 0.907216 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999998 +trainer/policy/normal/std Mean 2.79995 +trainer/policy/normal/std Std 0.68334 +trainer/policy/normal/std Max 5.6349 +trainer/policy/normal/std Min 0.277459 +trainer/policy/normal/log_std Mean 0.986614 +trainer/policy/normal/log_std Std 0.332377 +trainer/policy/normal/log_std Max 1.72898 +trainer/policy/normal/log_std Min -1.28208 +eval/num steps total 994320 +eval/num paths total 998 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.100231 +eval/Actions Std 0.902548 +eval/Actions Max 0.999998 +eval/Actions Min -0.999993 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.53692 +time/logging (s) 0.0037043 +time/sampling batch (s) 0.279936 +time/saving (s) 0.00352123 +time/training (s) 6.64586 +time/epoch (s) 9.46994 +time/total (s) 10133.3 +Epoch -3 +---------------------------------- --------------- +2022-05-10 15:59:54.631309 PDT | [0] Epoch -2 finished +---------------------------------- --------------- +epoch -2 +replay_buffer/size 999033 +trainer/num train calls 999000 +trainer/Policy Loss -20.7025 +trainer/Log Pis Mean 24.8548 +trainer/Log Pis Std 13.0943 +trainer/Log Pis Max 62.1488 +trainer/Log Pis Min -7.01184 +trainer/policy/mean Mean -0.0397151 +trainer/policy/mean Std 0.910468 +trainer/policy/mean Max 0.999992 +trainer/policy/mean Min -0.999999 +trainer/policy/normal/std Mean 2.83251 +trainer/policy/normal/std Std 0.66488 +trainer/policy/normal/std Max 5.98895 +trainer/policy/normal/std Min 0.229503 +trainer/policy/normal/log_std Mean 0.999994 +trainer/policy/normal/log_std Std 0.328772 +trainer/policy/normal/log_std Max 1.78992 +trainer/policy/normal/log_std Min -1.47184 +eval/num steps total 995320 +eval/num paths total 999 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.135728 +eval/Actions Std 0.853535 +eval/Actions Max 0.999996 +eval/Actions Min -0.999988 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.42361 +time/logging (s) 0.00399514 +time/sampling batch (s) 0.281744 +time/saving (s) 0.00431529 +time/training (s) 6.40021 +time/epoch (s) 9.11388 +time/total (s) 10142.4 +Epoch -2 +---------------------------------- --------------- +2022-05-10 16:00:04.637517 PDT | [0] Epoch -1 finished +---------------------------------- --------------- +epoch -1 +replay_buffer/size 999033 +trainer/num train calls 1e+06 +trainer/Policy Loss -20.5702 +trainer/Log Pis Mean 24.6431 +trainer/Log Pis Std 13.6535 +trainer/Log Pis Max 80.3527 +trainer/Log Pis Min -7.13538 +trainer/policy/mean Mean -0.038432 +trainer/policy/mean Std 0.9111 +trainer/policy/mean Max 0.99999 +trainer/policy/mean Min -0.999992 +trainer/policy/normal/std Mean 2.76992 +trainer/policy/normal/std Std 0.665228 +trainer/policy/normal/std Max 5.15398 +trainer/policy/normal/std Min 0.22516 +trainer/policy/normal/log_std Mean 0.975079 +trainer/policy/normal/log_std Std 0.339427 +trainer/policy/normal/log_std Max 1.63977 +trainer/policy/normal/log_std Min -1.49095 +eval/num steps total 996320 +eval/num paths total 1000 +eval/path length Mean 1000 +eval/path length Std 0 +eval/path length Max 1000 +eval/path length Min 1000 +eval/Rewards Mean 0 +eval/Rewards Std 0 +eval/Rewards Max 0 +eval/Rewards Min 0 +eval/Returns Mean 0 +eval/Returns Std 0 +eval/Returns Max 0 +eval/Returns Min 0 +eval/Actions Mean -0.0836077 +eval/Actions Std 0.904052 +eval/Actions Max 0.999994 +eval/Actions Min -0.999999 +eval/Num Paths 1 +eval/Average Returns 0 +eval/normalized_score 0 +time/evaluation sampling (s) 2.64582 +time/logging (s) 0.00382055 +time/sampling batch (s) 0.528598 +time/saving (s) 0.00645694 +time/training (s) 6.79653 +time/epoch (s) 9.98122 +time/total (s) 10152.4 +Epoch -1 +---------------------------------- ---------------