| !!python/object/apply:collections.OrderedDict |
| - - - batch_size |
| - 8 |
| - - clip_range |
| - 0.4 |
| - - ent_coef |
| - 0.00013057334805552262 |
| - - gae_lambda |
| - 0.92 |
| - - gamma |
| - 0.98 |
| - - learning_rate |
| - 3.791707778339674e-05 |
| - - max_grad_norm |
| - 0.6 |
| - - n_envs |
| - 1 |
| - - n_epochs |
| - 5 |
| - - n_steps |
| - 2048 |
| - - n_timesteps |
| - 1000000.0 |
| - - normalize |
| - gamma: 0.98 |
| norm_obs: false |
| norm_reward: true |
| - - policy |
| - MlpPolicy |
| - - policy_kwargs |
| - activation_fn: !!python/name:torch.nn.modules.activation.ReLU '' |
| features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor '' |
| net_arch: |
| - pi: |
| - 256 |
| - 256 |
| vf: |
| - 256 |
| - 256 |
| - - vf_coef |
| - 0.6167177795726859 |
|
|