Skip to main content

The OpenAI Gym: A toolkit for developing and comparing your reinforcement learning agents.

Project description

https://travis-ci.org/openai/gym.svg?branch=master

OpenAI Gym is a toolkit for developing and comparing reinforcement learning algorithms. This is the gym open-source library, which gives you access to an ever-growing variety of environments.

gym makes no assumptions about the structure of your agent, and is compatible with any numerical computation library, such as TensorFlow or Theano. You can use it from Python code, and soon from other languages.

If you’re not sure where to start, we recommend beginning with the docs on our site.

A whitepaper for OpenAI Gym is available at http://arxiv.org/abs/1606.01540, and here’s a BibTeX entry that you can use to cite it in a publication:

@misc{1606.01540,
        Author = {Greg Brockman and Vicki Cheung and Ludwig Pettersson and Jonas Schneider and John Schulman and Jie Tang and Wojciech Zaremba},
        Title = {OpenAI Gym},
        Year = {2016},
        Eprint = {arXiv:1606.01540},
}

Basics

There are two basic concepts in reinforcement learning: the environment (namely, the outside world) and the agent (namely, the algorithm you are writing). The agent sends actions to the environment, and the environment replies with observations and rewards (that is, a score).

The core gym interface is Env, which is the unified environment interface. There is no interface for agents; that part is left to you. The following are the Env methods you should know:

  • reset(self): Reset the environment’s state. Returns observation.

  • step(self, action): Step the environment by one timestep. Returns observation, reward, done, info.

  • render(self, mode=’human’, close=False): Render one frame of the environment. The default mode will do something human friendly, such as pop up a window. Passing the close flag signals the renderer to close any such windows.

Installation

You can perform a minimal install of gym with:

git clone https://github.com/openai/gym.git
cd gym
pip install -e .

If you prefer, you can do a minimal install of the packaged version directly from PyPI:

pip install gym

You’ll be able to run a few environments right away:

We recommend playing with those environments at first, and then later installing the dependencies for the remaining environments.

Installing everything

To install the full set of environments, you’ll need to have some system packages installed. We’ll build out the list here over time; please let us know what you end up installing on your platform.

On OSX:

brew install cmake boost boost-python sdl2 swig wget

On Ubuntu 14.04:

apt-get install -y python-numpy python-dev cmake zlib1g-dev libjpeg-dev xvfb libav-tools xorg-dev python-opengl libboost-all-dev libsdl2-dev swig

MuJoCo has a proprietary dependency we can’t set up for you. Follow the instructions in the mujoco-py package for help.

Once you’re ready to install everything, run pip install -e '.[all]' (or pip install 'gym[all]').

Supported systems

We currently support Linux and OS X running Python 2.7 or 3.5. Python 3 support should still be considered experimental – if you find any bugs, please report them!

In particular on OSX + Python3 you may need to run

brew install boost-python --with-python3

We will expand support to Windows based on demand. We will also soon ship a Docker container exposing the environments callable from any platform, for use with any non-Python framework, such as Torch.

Pip version

To run pip install -e '.[all]', you’ll need a semi-recent pip. Please make sure your pip is at least at version 1.5.0. You can upgrade using the following: pip install --ignore-installed pip. Alternatively, you can open setup.py and install the dependencies by hand.

Rendering on a server

If you’re trying to render video on a server, you’ll need to connect a fake display. The easiest way to do this is by running under xvfb-run (on Ubuntu, install the xvfb package):

xvfb-run -s "-screen 0 1400x900x24" bash

Installing dependencies for specific environments

If you’d like to install the dependencies for only specific environments, see setup.py. We maintain the lists of dependencies on a per-environment group basis.

Environments

The code for each environment group is housed in its own subdirectory gym/envs. The specification of each task is in gym/envs/__init__.py. It’s worth browsing through both.

Algorithmic

These are a variety of algorithmic tasks, such as learning to copy a sequence.

import gym
env = gym.make('Copy-v0')
env.reset()
env.render()

Atari

The Atari environments are a variety of Atari video games. If you didn’t do the full install, you can install dependencies via pip install -e '.[atari]' (you’ll need cmake installed) and then get started as follow:

import gym
env = gym.make('SpaceInvaders-v0')
env.reset()
env.render()

This will install atari-py, which automatically compiles the Arcade Learning Environment. This can take quite a while (a few minutes on a decent laptop), so just be prepared.

Board games

The board game environments are a variety of board games. If you didn’t do the full install, you can install dependencies via pip install -e '.[board_game]' (you’ll need cmake installed) and then get started as follow:

import gym
env = gym.make('Go9x9-v0')
env.reset()
env.render()

Box2d

Box2d is a 2D physics engine. You can install it via pip install -e '.[box2d]' and then get started as follow:

import gym
env = gym.make('LunarLander-v2')
env.reset()
env.render()

Classic control

These are a variety of classic control tasks, which would appear in a typical reinforcement learning textbook. If you didn’t do the full install, you will need to run pip install -e '.[classic_control]' to enable rendering. You can get started with them via:

import gym
env = gym.make('CartPole-v0')
env.reset()
env.render()

MuJoCo

MuJoCo is a physics engine which can do very detailed efficient simulations with contacts. It’s not open-source, so you’ll have to follow the instructions in mujoco-py to set it up. You’ll have to also run pip install -e '.[mujoco]' if you didn’t do the full install.

import gym
env = gym.make('Humanoid-v0')
env.reset()
env.render()

Toy text

Toy environments which are text-based. There’s no extra dependency to install, so to get started, you can just do:

import gym
env = gym.make('FrozenLake-v0')
env.reset()
env.render()

Examples

See the examples directory.

Testing

We are using nose2 for tests. You can run them via:

nose2

You can also run tests in a specific directory by using the -s option, or by passing in the specific name of the test. See the nose2 docs for more details.

What’s new

  • 2016-11-1: Several experimental changes to how a running monitor interacts with environments. The monitor will now raise an error if reset() is called when the env has not returned done=True. The monitor will only record complete episodes where done=True. Finally, the monitor no longer calls seed() on the underlying env, nor does it record or upload seed information.

  • 2016-10-31: We’re experimentally expanding the environment ID format to include an optional username.

  • 2016-09-21: Switch the Gym automated logger setup to configure the root logger rather than just the ‘gym’ logger.

  • 2016-08-17: Calling close on an env will also close the monitor and any rendering windows.

  • 2016-08-17: The monitor will no longer write manifest files in real-time, unless write_upon_reset=True is passed.

  • 2016-05-28: For controlled reproducibility, envs now support seeding (cf #91 and #135). The monitor records which seeds are used. We will soon add seed information to the display on the scoreboard.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gym-0.5.3.tar.gz (145.7 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page