Fully differentiable reinforcement learning environments, written in Ivy.
Project description
What is Ivy Gym?
Ivy Gym opens the door for intersectional research between supervised learning (SL), reinforcement learning (RL), and trajectory optimization (TO), by implementing RL environments in a fully differentiable manner.
Specifically, Ivy gym provides differentiable implementations of the classic control tasks from OpenAI Gym, as well
as a new Swimmer task, which illustrates the simplicity of creating new tasks using Ivy. The differentiable nature
of the environments means that the cumulative reward can be directly optimized for in a supervised manner, without
need for reinforcement learning, which is the de facto approach for optimizing cumulative rewards. Ivy currently
supports Jax, TensorFlow, PyTorch, MXNet and Numpy. Check out the [docs](https://ivy-dl.org/gym) for more info!
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
ivy-gym-1.1.5.tar.gz
(11.1 kB
view hashes)
Built Distribution
ivy_gym-1.1.5-py3-none-any.whl
(17.1 kB
view hashes)