Skip to main content

Dynamic RL Environments for Autonomous Driving and Robot Soccer

Project description

DynEnv

Dynamic Simulation Environments for Reinforcement Learning

This project contains two reinforcement learning environments based on 2D physics simulation via pymunk. The environments support different observation modalities and also noisy observations. The current environments are the following:

  • Robot Soccer SPL League (RoboCupEnvironment): Here, two teams of robots are competing to play soccer.
  • Autonomous driving environment (DrivingEnvironment): Here, two teams of cars try to get to their unique destinations as quickly as possible without crashing or hitting pedestrians. The teams are not competing here, but only cars on the same team are allowed to share information (to model human drivers).

Table of contents

Requirements

  • Python 3.6+
  • PyMunk
  • OpenCV
  • PyGame
  • PyTorch (optional)

Installation

You can install simply using pip:

pip install DynEnv

Or build from source:

git clone https://github.com/szemenyeim/DynEnv.git
cd DynEnv
pip install -e .

Usage

You can simply use the environments the following way:

from DynEnv import *

myEnv = RoboCupEnvironment(nPlayers)
myEnv = DrivingEnvironment(nPlayers)

ret = myEnv.step(actions)

Or create vectorized environments by using:

make_dyn_env(env, num_envs, num_players, render, observationType, noiseType, noiseMagnitude, use_continuous_actions)
env = DynEnvType.ROBO_CUP or DynEnvType.DRIVE

More complex examples including

  • neural networks tailored for the special output format (i.e. the number of observations can vary through time),
  • logging and
  • plotting the results.

For the above, confer the DynEnv/examples directory. The main.py file consists a full example, while if you would like to try out how the environments work by hand, play.py is there for you as well.

Model structure

The most important part from the point of view of the neural network is the DynEnv/models directory, which exposes you the following classes:

  • ICMAgent: the top-level agent consisting of an A2C and an Intrinsic Curiosity Module (and its variant, Rational Curiosity Module)
  • InOutArranger: helper class to rearrange observations for simple NN forwarding
  • EmbedBlock: the embedding network used for an object
  • InputLayer: a complex network which convert all observations into a unified feature space
  • ActorBlock: a neural network predicting actions for a given action type
  • ActorLayer: an ensemble of ActorBlock to predict every action
  • AttentionLayer:
  • DynEnvFeatureExtractor: a wrapper for the input transform by InputLayer, collapsing the time dimension with Recurrent Temporal Attention and running an LSTM

Parameters

Here are some of the important settings of the environments

  • nPlayers [1-5/10]: Number of total players in the environment (in the RoboCup env this is per team). The limit is 10 in the Driving, 5 in the RoboCup env.
  • render [bool]: Whether to visualize the environment.
  • observationType [Full, Partial, Image]: Image observation only supported for the RoboCup environment.
  • noiseType [Random, Realistic]: Realistic noise: noise magnitude and false negative rate depends on distance, proximity of other objects and sighting type. False positives and misclassifications are more likely to occur in certain situations.
  • noiseMagnitude [0-5]: Variable to control noise
  • continuousActions [bool]: Whether the driving env actions are understood as categorical or continuous. (Driving env only)
  • allowHeadturn [bool]: Enables head turining actions. (RoboCup env only)

Here are some examples of different noise and observation types

Random Noise

Full Observation Partial Observation
Top Camera Bottom Camera

Realistic Noise

Full Observation Partial Observation
Top Camera Bottom Camera

Large, Realistic Noise

Full Observation Partial Observation
Top Camera Bottom Camera

Driving, Realistic Noise

Full Observation Partial Observation

Important functions and members

  • reset() Resets the environment to a new game and returns initial observations.
  • setRandomSeed(seed) Sets the environment seed, resets the environment and returns initial observations.
  • observation_space Returns information about the observations returned by the environrment.
  • action_space Returns information about the actions the environment expects.
  • step(actions) Performs one step. This consists of several simulation steps (10 for the Driving and 50 for the RoboCup environments). It returns observations for every 10 simulation steps and full state for the last step.
  • renderMode Whether to render to a display ('human') or to a memory array ('memory').
  • agentVisID With this, you can visualize the observation of an agent during rendering.
  • render() Returns rendered images if the render mode is 'memory'. Does nothing otherwise, as the rendering is done by the step function due to the multi-timestep feature.

So, what are the actions?

The environments expect an iterable object containing the actions for every player. Each player action must contain the following:

RoboCup:

  • Movement direction: Categorical (5)
  • Turn: Categorical (3)
  • Turn head: Continuous [-6 +6]
  • Kick: Categorical (3) (this is exclusive with moving or turning)

Driving:

  • Gas/break: Continuous [-3 +3] or Categorical (3)
  • Turn: Continuous [-3 +3] or Categorical (3)

What is returned?

Both environments return the following variables in the step function:

  • Observations: Observations for every agent. What this is exactly depends on the observationType variable.
  • Rewards: Rewards for each agent.
    • Team rewards: Shared rewards for every team. These are added to the agent reward variables, and are not returned.
  • Finished: Game over flag
  • Info: Other important data
    • Full State: The full state of the env
    • episode_r: Cumulative rewards for the episode (Returned only at the end of an episode)
    • episode_p_r: Cumulative positive-only rewards for the episode (Returned only at the end of an episode)
    • episode_g: Goals in these episode. For RoboCup this is goals per team, for the Driving env the first value is the number of cars that reached their destination without crashing, the second is the number of crashed cars. (Returned only at the end of an episode)

Position information is normalized in both the observations and the full state.

The Observation Space

Due to limitations in the OpenAI gym, this part of the environment is not fully compatible. The observation_space variable is an instance of gym.space.Space, however, the meaning is slightly different. I.e. querying the observation_space variable and calling .sample() on it will get you a fully valid observation format, it does not cover every form of observations an environment can produce. Let us elaborate on that!

Gym-like observation space descriptor

Due to the fact that in every time step each agent can see different number of objects (such as cars in the Driving environment), including 0 as a valid number for each object type (not to mention false positive sightings or misclassifications), we cannot give an observation space format which covers all possibilities. However, what we can do is to assume that each object type is present in the observation with a single instance, thus including every necessary information about the object space (but be aware that multiple observations from the same object type can be in the list of observations).

Here is an example for the Driving environment how the observation space looks like (we use extensively the Dict gym space, as it enables to describe what is contained, note that prefixes are present as for our network the order in the featuresPerObject - see below example - matters):

...
# subspace for cars
 car_space = Dict({
            "position": Box(-self.mean * 2, +self.mean * 2, shape=(2,)),
            "orientation": Box(-1, 1, shape=(2,)),
            "width_height": Box(-10, 10, shape=(2,)),
            "finished": MultiBinary(1)
        })
...

# assemble observation space
self.observation_space = Dict({
                "0_self": self_space,
                "1_car": car_space,
                "2_obstacle": obstacle_space,
                "3_pedestrian": pedestrian_space,
                "4_lane": lane_space
            })
List of observations

The observations returned are arranged as follows:

[nParallelEnvs x nTimeSteps x nAgents x nObjectType]

Each element of the above list is a NumPy array containing all the observations by a single agent in a single timestep. To help contructing input layers a custom class DynEnv.models.InOutArranger is provided with the following two functions:

  • inputs, counts = rearrange_inputs(x): Creates a single list of NumPy arrays. Each element of this list contains a single numpy array of all the observations for a given object type. (Warning: in some cases this might be an empty list!)
  • outputs, masks = rearrange_outputs(inputs, counts, device): Takes a list of Torch Tensors and the counts output by the previous function, and creates a single tensor shaped [TimeSteps x maxObjCnt x nPlayers x featureCnt] by padding the second dimension to the largest number of objects seen for every robot. The masks variable is binary array shaped [TimeSteps x maxObjCnt x nPlayers], which is True for padded elements (this is in line with PyTorch's MultiHeadedAttention layer). (Warning: This assumes that the featureCnt is the same for every object time.)

Here is a more comprehensive example:

from DynEnv.models import *
from torch import nn

# setup environment, query all required variables
myEnv = ...
obsSpace = myEnv.observation_space
nTime =  5 if env is DynEnvType.ROBO_CUP else 1
nPlayers = ...
nObjectTypes = len(obsSpace.spaces.keys())
featuresPerObject = [flatdim(s) for s in obsSpace.spaces.values()]

# create neural network and rearrange inputs
device = <CUDA or CPU>
myNeuralNets = [nn.Linear(objfeat,128).to(device) for objFeat in featuresPerObject]
myArranger = models.InOutArranger(nObjectTypes,nPlayers,nTime)

...
# create sample action and step
actions = torch.stack([action_space.sample() for _ in range(nPlayers)]
obs, _ = myEnv.step(actions)

# summary
# rearrange inputs - forward - rearrange outputs
netInputs, counts = myArranger.rearrange_inputs(obs)
netOutputs = [myNet(torch.tensor(netInput).to(device)) for myNet,netInput in zip(myNeuralNets,netInputs)]
outputs,masks = myArranger.rearrange_outputs(netOutputs,counts,device)

RoboCup

The full state contains the following:

  • Robots [x, y, cos(angle), sin(angle), team ID, fallen or penalized]
  • Balls [x, y, ball owned by team ID, closest robot status]

'Team ID' is +/-1. 'Fallen or penalized' and 'closest robot status' are binary numbers. The latter is 1 for the robot closest to the ball from each team.

If the observation is full state, the robot's own position is returned in a separate list, and both axes are flipped and angles rotated 180 degrees for team -1. Moreover, in this case the ball owned flag indicates whether the ball is owned by the robot's team, or the opponent.

The partial observation contains the following for each robot:

  • Balls: [x, y, radius, ball owned status, closest robot status]
  • Robots (self not included): [x, y, cos(angle), sin(angle), team, fallen or penalized]
  • Goalposts: [x, y, radius]
  • Crosses: [x, y, radius]
  • Lines: [x1, y1, x2, y2]
  • Center circle: [x, y, radius]

Ball owned status is 0 if the ball is not owned, +1 if the ball is owned by the robot's team and -1 if owned by the opposite team.

In the partial sighting case, the positions and angles are returned relative to the robot's position and head angle.

The image observations contain 2D images of semantic labels. The images have 4 binary channels:

  • 0: Ball
  • 1: Robot
  • 2: Goalpost
  • 3: Line

Driving

The full state contains the following:

  • Cars: [x, y, cos(angle), sin(angle), width, height, finished]
  • Obstacles: [x, y, cos(angle), sin(angle), width, height]
  • Pedestrians: [x, y]
  • Lanes: [x1, y1, x2, y2, type]

Lane type is 0 for standard lanes, 1 for the middle lane and -1 for the edge of the road.

If the observation is full state, the car's own position is returned in a separate list, identical to the Self entry below.

The partial observation contains the following for each car:

  • Self: [x, y, cos(angle), sin(angle), width, height, goal_x, goal_y, finished]
  • Cars: [x, y, cos(angle), sin(angle), width, height]
  • Obstacles: [x, y, cos(angle), sin(angle), width, height]
  • Pedestrians: [x, y]
  • Lanes: [signed distance, cos(angle), sin(angle), type]

Widths and heights are also normalized.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Files for DynEnv, version 1.1.1
Filename, size File type Python version Upload date Hashes
Filename, size DynEnv-1.1.1-py3-none-any.whl (87.7 kB) File type Wheel Python version py3 Upload date Hashes View hashes
Filename, size DynEnv-1.1.1.tar.gz (66.6 kB) File type Source Python version None Upload date Hashes View hashes

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN SignalFx SignalFx Supporter DigiCert DigiCert EV certificate StatusPage StatusPage Status page