| !!python/object/apply:collections.OrderedDict |
| - - - batch_size |
| - 16 |
| - - clip_range |
| - 0.3 |
| - - ent_coef |
| - 3.1441389214159857e-06 |
| - - gae_lambda |
| - 0.8 |
| - - gamma |
| - 0.995 |
| - - learning_rate |
| - 0.00017959211641976886 |
| - - max_grad_norm |
| - 0.9 |
| - - n_epochs |
| - 10 |
| - - n_steps |
| - 2048 |
| - - n_timesteps |
| - 1000000.0 |
| - - normalize |
| - gamma: 0.995 |
| norm_obs: false |
| norm_reward: true |
| - - policy |
| - MlpPolicy |
| - - policy_kwargs |
| - activation_fn: !!python/name:torch.nn.modules.activation.Tanh '' |
| features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor '' |
| net_arch: |
| - pi: |
| - 64 |
| - 64 |
| vf: |
| - 64 |
| - 64 |
| - - vf_coef |
| - 0.4351450387648799 |
|
|