| !!python/object/apply:collections.OrderedDict |
| - - - batch_size |
| - 8 |
| - - clip_range |
| - 0.1 |
| - - ent_coef |
| - 5.167107294612664e-08 |
| - - gae_lambda |
| - 0.95 |
| - - gamma |
| - 0.999 |
| - - learning_rate |
| - 0.0001214437022727675 |
| - - max_grad_norm |
| - 2 |
| - - n_epochs |
| - 20 |
| - - n_steps |
| - 2048 |
| - - n_timesteps |
| - 1000000.0 |
| - - normalize |
| - gamma: 0.999 |
| norm_obs: false |
| norm_reward: true |
| - - policy |
| - MlpPolicy |
| - - policy_kwargs |
| - activation_fn: !!python/name:torch.nn.modules.activation.Tanh '' |
| features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor '' |
| net_arch: |
| - pi: |
| - 64 |
| - 64 |
| vf: |
| - 64 |
| - 64 |
| - - vf_coef |
| - 0.6162112311062333 |
|
|