.. _a2c: .. automodule:: stable_baselines3.a2c A2C ==== A synchronous, deterministic variant of `Asynchronous Advantage Actor Critic (A3C) `_. It uses multiple workers to avoid the use of a replay buffer. .. warning:: If you find training unstable or want to match performance of stable-baselines A2C, consider using ``RMSpropTFLike`` optimizer from ``stable_baselines3.common.sb2_compat.rmsprop_tf_like``. You can change optimizer with ``A2C(policy_kwargs=dict(optimizer_class=RMSpropTFLike, eps=1e-5))``. Read more `here `_. Notes ----- - Original paper: https://arxiv.org/abs/1602.01783 - OpenAI blog post: https://openai.com/blog/baselines-acktr-a2c/ Can I use? ---------- - Recurrent policies: ✔️ - Multi processing: ✔️ - Gym spaces: ============= ====== =========== Space Action Observation ============= ====== =========== Discrete ✔️ ✔️ Box ✔️ ✔️ MultiDiscrete ✔️ ✔️ MultiBinary ✔️ ✔️ ============= ====== =========== Example ------- Train a A2C agent on ``CartPole-v1`` using 4 environments. .. code-block:: python import gym from stable_baselines3 import A2C from stable_baselines3.a2c import MlpPolicy from stable_baselines3.common.env_util import make_vec_env # Parallel environments env = make_vec_env('CartPole-v1', n_envs=4) model = A2C(MlpPolicy, env, verbose=1) model.learn(total_timesteps=25000) model.save("a2c_cartpole") del model # remove to demonstrate saving and loading model = A2C.load("a2c_cartpole") obs = env.reset() while True: action, _states = model.predict(obs) obs, rewards, dones, info = env.step(action) env.render() Results ------- Atari Games ^^^^^^^^^^^ The complete learning curves are available in the `associated PR #110 `_. PyBullet Environments ^^^^^^^^^^^^^^^^^^^^^ Results on the PyBullet benchmark (2M steps) using 6 seeds. The complete learning curves are available in the `associated issue #48 `_. .. note:: Hyperparameters from the `gSDE paper `_ were used (as they are tuned for PyBullet envs). *Gaussian* means that the unstructured Gaussian noise is used for exploration, *gSDE* (generalized State-Dependent Exploration) is used otherwise. +--------------+--------------+--------------+--------------+-------------+ | Environments | A2C | A2C | PPO | PPO | +==============+==============+==============+==============+=============+ | | Gaussian | gSDE | Gaussian | gSDE | +--------------+--------------+--------------+--------------+-------------+ | HalfCheetah | 2003 +/- 54 | 2032 +/- 122 | 1976 +/- 479 | 2826 +/- 45 | +--------------+--------------+--------------+--------------+-------------+ | Ant | 2286 +/- 72 | 2443 +/- 89 | 2364 +/- 120 | 2782 +/- 76 | +--------------+--------------+--------------+--------------+-------------+ | Hopper | 1627 +/- 158 | 1561 +/- 220 | 1567 +/- 339 | 2512 +/- 21 | +--------------+--------------+--------------+--------------+-------------+ | Walker2D | 577 +/- 65 | 839 +/- 56 | 1230 +/- 147 | 2019 +/- 64 | +--------------+--------------+--------------+--------------+-------------+ How to replicate the results? ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Clone the `rl-zoo repo `_: .. code-block:: bash git clone https://github.com/DLR-RM/rl-baselines3-zoo cd rl-baselines3-zoo/ Run the benchmark (replace ``$ENV_ID`` by the envs mentioned above): .. code-block:: bash python train.py --algo a2c --env $ENV_ID --eval-episodes 10 --eval-freq 10000 Plot the results (here for PyBullet envs only): .. code-block:: bash python scripts/all_plots.py -a a2c -e HalfCheetah Ant Hopper Walker2D -f logs/ -o logs/a2c_results python scripts/plot_from_file.py -i logs/a2c_results.pkl -latex -l A2C Parameters ---------- .. autoclass:: A2C :members: :inherited-members: A2C Policies ------------- .. autoclass:: MlpPolicy :members: :inherited-members: .. autoclass:: stable_baselines3.common.policies.ActorCriticPolicy :members: :noindex: .. autoclass:: CnnPolicy :members: .. autoclass:: stable_baselines3.common.policies.ActorCriticCnnPolicy :members: :noindex: