Skip to main content

A reinforcement learning environment for simulating a plane in 2D available, with gymnasium and gymnax support

Project description

✈️ Plane: Reinforcement Learning Environment for Aircraft Control

Demo of Plane environment

Plane is a lightweight yet realistic reinforcement learning environment simulating a 2D side view of an Airbus A320-like aircraft. It’s designed for fast, end-to-end training on GPU with JAX while staying physics-based and realistic enough to capture the core challenges of aircraft control.

Plane allows you to benchmark RL agents on delays, irrecoverable states, partial observability, and competing objectives — challenges that are often ignored in standard toy environments.


✨ Features

  • 🏎 Fast & parallelizable thanks to JAX — scale to thousands of parallel environments on GPU/TPU.
  • 📐 Physics-based: Dynamics are derived from airplane modeling equations (not arcade physics).
  • 🧪 Reliable: Covered by unit tests to ensure stability and reproducibility.
  • 🎯 Challenging: Captures real-world aviation control problems (momentum, delays, irrecoverable states).
  • 🔄 Compatible with multiple interfaces: Designed to work with JAX-based environments.
  • 🌟 Upcoming features: Environmental perturbations (e.g., wind) will be available in future releases.

📊 Stable Altitude vs. Power & Pitch

Below is an example of how stable altitude changes with engine power and pitch:

Stable altitude graph

This highlights the multi-stability phenomenon: holding a constant power setting can lead the plane to naturally converge to a stable altitude.


🚀 Installation

Once released on PyPI, you can install Plane with:

# Using pip
pip install plane-env

# Or with Poetry
poetry add plane-env

🎮 Usage

Here’s a minimal example of running an episode and saving a video:

from plane_env.env_jax import Airplane2D, EnvParams

# Create env
env = Airplane2D()
seed = 42
env_params = EnvParams(max_steps_in_episode=1_000)

# Simple constant policy with 80% power and 0° stick input.
action = (0.8, 0.0)

# Save the video
env.save_video(lambda o: action, seed, folder="videos", episode_index=0, params=env_params, format="gif")

Of course you can also directly use it to train an agent using your favorite RL library (here: stable-baselines3)

from plane_env.env_gymnasium import Airplane2D, EnvParams
from stable_baselines3 import SAC

# Create env
env = Airplane2D()
# Model training (adapted from https://stable-baselines3.readthedocs.io/en/master/modules/sac.html)


model = SAC("MlpPolicy", env, verbose=1)
model.learn(total_timesteps=10_000, log_interval=4)
model.save("sac_plane")

del model # remove to demonstrate saving and loading

model = SAC.load("sac_plane")

obs, info = env.reset()
while True:
    action, _states = model.predict(obs, deterministic=True)
    obs, reward, terminated, truncated, info = env.step(action)
    if terminated or truncated:
        break

🛩️ Environment Overview (Reinforcement Learning Perspective)

State (EnvState): 13-dimensional vector representing aircraft dynamics:

Variable Description
x Horizontal position (m)
x_dot Horizontal speed (m/s)
z Altitude (m)
z_dot Vertical speed (m/s)
theta Pitch angle (rad)
theta_dot Pitch angular velocity (rad/s)
alpha Angle of attack (rad)
gamma Flight path angle (rad)
m Aircraft mass (kg)
power Normalized engine thrust (0–1)
stick Control stick input for pitch (–1 to 1)
fuel Remaining fuel (kg)
t Current timestep
target_altitude Desired target altitude (m)

The state also provides derived properties like air density, Mach number, and speed of sound.

The agent currently observes all of the state, minus x and t (as they should be irrelevant for control), as well as fuel which is currently not used.

Action Space: Continuous 2D vector [power_requested, stick_requested] controlling engine thrust and pitch.

Reward Function:

  • Encourages maintaining target altitude.
  • Terminal altitude violations (z < min_alt or z > max_alt) incur -max_steps_in_episode.
  • Otherwise, reward is sthe quared normalized difference to target altitude:

$r_t = \left( \frac{\text{max\_alt} - | \text{target\_altitude} - z_t |}{\text{max\_alt} - \text{min\_alt}} \right)^2$

Episode Termination:

  • Altitude limits exceeded → terminated
  • Maximum episode length reached → truncated

Time step: delta_t = 0.5 s, max_steps_in_episode = 1,000.


🧩 Challenges Modeled

Plane is designed to test RL agents under realistic aviation challenges:

  • Delay: Engine power changes take time to fully apply.
  • 👀 Partial observability: Some forces cannot be directly measured.
  • 🏁 Competing objectives: Reach target altitude fast while minimizing fuel and overshoot.
  • 🌀 Momentum effects: Control inputs show delayed impact due to physical inertia.
  • ⚠️ Irrecoverable states: Certain trajectories inevitably lead to failure (crash).

Environmental perturbations (wind, turbulence) are coming in a future release.


📦 Roadmap

  • Add perturbations (wind with varying speeds and directions) to model the non-stationarity of the dynamics.
  • Add an easier interface to create partially-observable versions of the environment.
  • Provide ready-to-use benchmark results for popular RL baselines.
  • Add fuel consumption.

🤝 Contributing

Contributions are welcome! Please open an issue or PR if you have suggestions, bug reports, or new features.


📜 License

MIT License – feel free to use it in your own research and projects.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

plane_env-0.1.1.1.tar.gz (24.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

plane_env-0.1.1.1-py3-none-any.whl (24.9 kB view details)

Uploaded Python 3

File details

Details for the file plane_env-0.1.1.1.tar.gz.

File metadata

  • Download URL: plane_env-0.1.1.1.tar.gz
  • Upload date:
  • Size: 24.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.7.1 CPython/3.12.11 Linux/6.11.0-1018-azure

File hashes

Hashes for plane_env-0.1.1.1.tar.gz
Algorithm Hash digest
SHA256 bc7407a053699be0bf8289da0c538b4501151ff794226e838c3d1adb15ef627d
MD5 e8e55b6851411ca8a6eff94888f4e259
BLAKE2b-256 9dbe88ca51bb655e663b02e1cb841726bdd5c63d6ebf4b6a17e4657772ec99b8

See more details on using hashes here.

File details

Details for the file plane_env-0.1.1.1-py3-none-any.whl.

File metadata

  • Download URL: plane_env-0.1.1.1-py3-none-any.whl
  • Upload date:
  • Size: 24.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.7.1 CPython/3.12.11 Linux/6.11.0-1018-azure

File hashes

Hashes for plane_env-0.1.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 32640f036d9746b90c9cce5204af6dee3c3a1b0ad54d6cc04301a1255ad27e50
MD5 af543313720b45a1207467323e4e5333
BLAKE2b-256 55b62ea25cf98cceae77aee94c6ce4b474711d371b9e52fa3e04940c39fe1b6c

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page