| # Table of Contents |
|
|
| * [mlagents\_envs.envs.unity\_gym\_env](#mlagents_envs.envs.unity_gym_env) |
| * [UnityGymException](#mlagents_envs.envs.unity_gym_env.UnityGymException) |
| * [UnityToGymWrapper](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper) |
| * [\_\_init\_\_](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.__init__) |
| * [reset](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.reset) |
| * [step](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.step) |
| * [render](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.render) |
| * [close](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.close) |
| * [seed](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.seed) |
| * [ActionFlattener](#mlagents_envs.envs.unity_gym_env.ActionFlattener) |
| * [\_\_init\_\_](#mlagents_envs.envs.unity_gym_env.ActionFlattener.__init__) |
| * [lookup\_action](#mlagents_envs.envs.unity_gym_env.ActionFlattener.lookup_action) |
|
|
| <a name="mlagents_envs.envs.unity_gym_env"></a> |
| # mlagents\_envs.envs.unity\_gym\_env |
| |
| <a name="mlagents_envs.envs.unity_gym_env.UnityGymException"></a> |
| ## UnityGymException Objects |
| |
| ```python |
| class UnityGymException(error.Error) |
| ``` |
| |
| Any error related to the gym wrapper of ml-agents. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper"></a> |
| ## UnityToGymWrapper Objects |
| |
| ```python |
| class UnityToGymWrapper(gym.Env) |
| ``` |
| |
| Provides Gym wrapper for Unity Learning Environments. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.__init__"></a> |
| #### \_\_init\_\_ |
|
|
| ```python |
| | __init__(unity_env: BaseEnv, uint8_visual: bool = False, flatten_branched: bool = False, allow_multiple_obs: bool = False, action_space_seed: Optional[int] = None) |
| ``` |
|
|
| Environment initialization |
|
|
| **Arguments**: |
|
|
| - `unity_env`: The Unity BaseEnv to be wrapped in the gym. Will be closed when the UnityToGymWrapper closes. |
| - `uint8_visual`: Return visual observations as uint8 (0-255) matrices instead of float (0.0-1.0). |
| - `flatten_branched`: If True, turn branched discrete action spaces into a Discrete space rather than |
| MultiDiscrete. |
| - `allow_multiple_obs`: If True, return a list of np.ndarrays as observations with the first elements |
| containing the visual observations and the last element containing the array of vector observations. |
| If False, returns a single np.ndarray containing either only a single visual observation or the array of |
| vector observations. |
| - `action_space_seed`: If non-None, will be used to set the random seed on created gym.Space instances. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.reset"></a> |
| #### reset |
|
|
| ```python |
| | reset() -> Union[List[np.ndarray], np.ndarray] |
| ``` |
|
|
| Resets the state of the environment and returns an initial observation. |
| Returns: observation (object/list): the initial observation of the |
| space. |
|
|
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.step"></a> |
| #### step |
|
|
| ```python |
| | step(action: List[Any]) -> GymStepResult |
| ``` |
|
|
| Run one timestep of the environment's dynamics. When end of |
| episode is reached, you are responsible for calling `reset()` |
| to reset this environment's state. |
| Accepts an action and returns a tuple (observation, reward, done, info). |
|
|
| **Arguments**: |
|
|
| - `action` _object/list_ - an action provided by the environment |
|
|
| **Returns**: |
|
|
| - `observation` _object/list_ - agent's observation of the current environment |
| reward (float/list) : amount of reward returned after previous action |
| - `done` _boolean/list_ - whether the episode has ended. |
| - `info` _dict_ - contains auxiliary diagnostic information. |
|
|
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.render"></a> |
| #### render |
|
|
| ```python |
| | render(mode="rgb_array") |
| ``` |
|
|
| Return the latest visual observations. |
| Note that it will not render a new frame of the environment. |
|
|
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.close"></a> |
| #### close |
|
|
| ```python |
| | close() -> None |
| ``` |
|
|
| Override _close in your subclass to perform any necessary cleanup. |
| Environments will automatically close() themselves when |
| garbage collected or when the program exits. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.seed"></a> |
| #### seed |
| |
| ```python |
| | seed(seed: Any = None) -> None |
| ``` |
| |
| Sets the seed for this env's random number generator(s). |
| Currently not implemented. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.ActionFlattener"></a> |
| ## ActionFlattener Objects |
| |
| ```python |
| class ActionFlattener() |
| ``` |
| |
| Flattens branched discrete action spaces into single-branch discrete action spaces. |
| |
| <a name="mlagents_envs.envs.unity_gym_env.ActionFlattener.__init__"></a> |
| #### \_\_init\_\_ |
|
|
| ```python |
| | __init__(branched_action_space) |
| ``` |
|
|
| Initialize the flattener. |
|
|
| **Arguments**: |
|
|
| - `branched_action_space`: A List containing the sizes of each branch of the action |
| space, e.g. [2,3,3] for three branches with size 2, 3, and 3 respectively. |
|
|
| <a name="mlagents_envs.envs.unity_gym_env.ActionFlattener.lookup_action"></a> |
| #### lookup\_action |
| |
| ```python |
| | lookup_action(action) |
| ``` |
| |
| Convert a scalar discrete action into a unique set of branched actions. |
| |
| **Arguments**: |
| |
| - `action`: A scalar value representing one of the discrete actions. |
| |
| **Returns**: |
| |
| The List containing the branched actions. |
| |