| from typing import Optional, Union |
|
|
| import numpy as np |
| from gym import Env, Space |
| from gym.spaces import Box, Discrete, MultiBinary, MultiDiscrete |
|
|
| from stable_baselines3.common.type_aliases import GymObs, GymStepReturn |
|
|
|
|
| class IdentityEnv(Env): |
| def __init__(self, dim: Optional[int] = None, space: Optional[Space] = None, ep_length: int = 100): |
| """ |
| Identity environment for testing purposes |
| |
| :param dim: the size of the action and observation dimension you want |
| to learn. Provide at most one of ``dim`` and ``space``. If both are |
| None, then initialization proceeds with ``dim=1`` and ``space=None``. |
| :param space: the action and observation space. Provide at most one of |
| ``dim`` and ``space``. |
| :param ep_length: the length of each episode in timesteps |
| """ |
| if space is None: |
| if dim is None: |
| dim = 1 |
| space = Discrete(dim) |
| else: |
| assert dim is None, "arguments for both 'dim' and 'space' provided: at most one allowed" |
|
|
| self.action_space = self.observation_space = space |
| self.ep_length = ep_length |
| self.current_step = 0 |
| self.num_resets = -1 |
| self.reset() |
|
|
| def reset(self) -> GymObs: |
| self.current_step = 0 |
| self.num_resets += 1 |
| self._choose_next_state() |
| return self.state |
|
|
| def step(self, action: Union[int, np.ndarray]) -> GymStepReturn: |
| reward = self._get_reward(action) |
| self._choose_next_state() |
| self.current_step += 1 |
| done = self.current_step >= self.ep_length |
| return self.state, reward, done, {} |
|
|
| def _choose_next_state(self) -> None: |
| self.state = self.action_space.sample() |
|
|
| def _get_reward(self, action: Union[int, np.ndarray]) -> float: |
| return 1.0 if np.all(self.state == action) else 0.0 |
|
|
| def render(self, mode: str = "human") -> None: |
| pass |
|
|
|
|
| class IdentityEnvBox(IdentityEnv): |
| def __init__(self, low: float = -1.0, high: float = 1.0, eps: float = 0.05, ep_length: int = 100): |
| """ |
| Identity environment for testing purposes |
| |
| :param low: the lower bound of the box dim |
| :param high: the upper bound of the box dim |
| :param eps: the epsilon bound for correct value |
| :param ep_length: the length of each episode in timesteps |
| """ |
| space = Box(low=low, high=high, shape=(1,), dtype=np.float32) |
| super().__init__(ep_length=ep_length, space=space) |
| self.eps = eps |
|
|
| def step(self, action: np.ndarray) -> GymStepReturn: |
| reward = self._get_reward(action) |
| self._choose_next_state() |
| self.current_step += 1 |
| done = self.current_step >= self.ep_length |
| return self.state, reward, done, {} |
|
|
| def _get_reward(self, action: np.ndarray) -> float: |
| return 1.0 if (self.state - self.eps) <= action <= (self.state + self.eps) else 0.0 |
|
|
|
|
| class IdentityEnvMultiDiscrete(IdentityEnv): |
| def __init__(self, dim: int = 1, ep_length: int = 100): |
| """ |
| Identity environment for testing purposes |
| |
| :param dim: the size of the dimensions you want to learn |
| :param ep_length: the length of each episode in timesteps |
| """ |
| space = MultiDiscrete([dim, dim]) |
| super().__init__(ep_length=ep_length, space=space) |
|
|
|
|
| class IdentityEnvMultiBinary(IdentityEnv): |
| def __init__(self, dim: int = 1, ep_length: int = 100): |
| """ |
| Identity environment for testing purposes |
| |
| :param dim: the size of the dimensions you want to learn |
| :param ep_length: the length of each episode in timesteps |
| """ |
| space = MultiBinary(dim) |
| super().__init__(ep_length=ep_length, space=space) |
|
|
|
|
| class FakeImageEnv(Env): |
| """ |
| Fake image environment for testing purposes, it mimics Atari games. |
| |
| :param action_dim: Number of discrete actions |
| :param screen_height: Height of the image |
| :param screen_width: Width of the image |
| :param n_channels: Number of color channels |
| :param discrete: Create discrete action space instead of continuous |
| :param channel_first: Put channels on first axis instead of last |
| """ |
|
|
| def __init__( |
| self, |
| action_dim: int = 6, |
| screen_height: int = 84, |
| screen_width: int = 84, |
| n_channels: int = 1, |
| discrete: bool = True, |
| channel_first: bool = False, |
| ): |
| self.observation_shape = (screen_height, screen_width, n_channels) |
| if channel_first: |
| self.observation_shape = (n_channels, screen_height, screen_width) |
| self.observation_space = Box(low=0, high=255, shape=self.observation_shape, dtype=np.uint8) |
| if discrete: |
| self.action_space = Discrete(action_dim) |
| else: |
| self.action_space = Box(low=-1, high=1, shape=(5,), dtype=np.float32) |
| self.ep_length = 10 |
| self.current_step = 0 |
|
|
| def reset(self) -> np.ndarray: |
| self.current_step = 0 |
| return self.observation_space.sample() |
|
|
| def step(self, action: Union[np.ndarray, int]) -> GymStepReturn: |
| reward = 0.0 |
| self.current_step += 1 |
| done = self.current_step >= self.ep_length |
| return self.observation_space.sample(), reward, done, {} |
|
|
| def render(self, mode: str = "human") -> None: |
| pass |
|
|