Third-Party Environments#
The Farama Foundation maintains a number of other projects, most of which use Gymnasium. Topics include: multi-agent RL (PettingZoo), offline-RL (Minari), gridworlds (Minigrid), robotics (Gymnasium-Robotics), multi-objective RL (MO-Gymnasium) many-agent RL (MAgent2), 3D navigation (Miniworld), and many more.
This page contains environments which are not maintained by Farama Foundation and, as such, cannot be guaranteed to function as intended.
If you’d like to contribute an environment, please reach out on Discord.
highway-env: Autonomous driving and tactical decision-making tasks#
An environment for behavioral planning in autonomous driving, with an emphasis on high-level perception and decision rather than low-level sensing and control.
sumo-rl: Reinforcement Learning using SUMO traffic simulator#
Gymnasium wrapper for various environments in the SUMO traffic simulator. Supports both single and multiagent settings (using pettingzoo).
panda-gym: Robotics environments using the PyBullet physics engine#
PyBullet based simulations of a robotic arm moving objects.
tmrl: TrackMania 2020 through RL#
tmrl is a distributed framework for training Deep Reinforcement Learning AIs in real-time applications. It is demonstrated on the TrackMania 2020 video game.
gym-jiminy: Training Robots in Jiminy#
gym-jiminy presents an extension of the initial Gym for robotics using Jiminy, an extremely fast and light-weight simulator for poly-articulated systems using Pinocchio for physics evaluation and Meshcat for web-based 3D rendering.
Safety-Gymnasium: Ensuring safety in real-world RL scenarios#
Highly scalable and customizable Safe Reinforcement Learning library.
stable-retro: Classic retro games, a maintained version of OpenAI Retro#
Supported fork of gym-retro: turn classic video games into Gymnasium environments.
flappy-bird-gymnasium: A Flappy Bird environment for Gymnasium#
A simple environment for single-agent reinforcement learning algorithms on a clone of Flappy Bird, the hugely popular arcade-style mobile game. Both state and pixel observation environments are available.
gym-saturation: Environments used to prove theorems#
An environment for guiding automated theorem provers based on saturation algorithms (e.g. Vampire).
matrix-mdp: Easily create discrete MDPs#
An environment to easily implement discrete MDPs as gym environments. Turn a set of matrices (P_0(s)
, P(s'| s, a)
and R(s', s, a)
) into a gym environment that represents the discrete MDP ruled by these dynamics.
Third-Party Environments using Gym#
There are a large number of third-party environments using various versions of Gym. Many of these can be adapted to work with gymnasium (see Compatibility with Gym), but are not guaranteed to be fully functional.
Video Game environments#
gym-derk: GPU accelerated MOBA environment#
A 3v3 MOBA environment where you train creatures to fight each other.
SlimeVolleyGym: A simple environment for Slime Volleyball game#
A simple environment for benchmarking single and multi-agent reinforcement learning algorithms on a clone of Slime Volleyball game.
Unity ML Agents: Environments for Unity game engine#
Gym (and PettingZoo) wrappers for arbitrary and premade environments with the Unity game engine.
PGE: Parallel Game Engine#
Uses The Open 3D Engine for AI simulations and can interoperate with the Gym. Uses PyBullet physics.
Robotics environments#
gym-pybullet-drones: Environments for quadcopter control#
A simple environment using PyBullet to simulate the dynamics of a Bitcraze Crazyflie 2.x nanoquadrotor.
MarsExplorer: Environments for controlling robot on Mars#
Mars Explorer is a Gym compatible environment designed and developed as an initial endeavor to bridge the gap between powerful Deep Reinforcement Learning methodologies and the problem of exploration/coverage of an unknown terrain.
robo-gym: Real-world and simulation robotics#
Robo-gym provides a collection of reinforcement learning environments involving robotic tasks applicable in both simulation and real-world robotics.
Offworld-gym: Control real robots remotely for free#
Gym environments that let you control real robots in a laboratory via the internet.
safe-control-gym: Evaluate safety of RL algorithms#
Evaluate safety, robustness and generalization via PyBullet based CartPole and Quadrotor environments—with CasADi (symbolic) a priori dynamics and constraints.
gym-softrobot: Soft-robotics environments#
A large-scale benchmark for co-optimizing the design and control of soft robots.
iGibson: Photorealistic and interactive robotics environments#
A simulation environment with high-quality realistic scenes, with interactive physics using PyBullet.
DexterousHands: Dual dexterous hand manipulation tasks#
This is a library that provides dual dexterous hand manipulation tasks through Isaac Gym.
OmniIsaacGymEnvs: Gym environments for NVIDIA Omniverse Isaac #
Reinforcement Learning Environments for Omniverse Isaac simulator.
Autonomous Driving environments#
gym-duckietown: Lane-following for self-driving cars#
A lane-following simulator built for the Duckietown project (small-scale self-driving car course).
gym-electric-motor: Gym environments for electric motor simulations#
An environment for simulating a wide variety of electric drives taking into account different types of electric motors and converters.
CommonRoad-RL: Motion planning for traffic scenarios #
A Gym for solving motion planning problems for various traffic scenarios compatible with CommonRoad benchmarks, which provides configurable rewards, action spaces, and observation spaces.
racing_dreamer: Latent imagination in autonomous racing#
Train a model-based RL agent in simulation and, without finetuning, transfer it to small-scale race cars.
l2r: Multimodal control environment where agents learn how to race#
An open-source reinforcement learning environment for autonomous racing.
racecar_gym: Miniature racecar env using PyBullet#
A gym environment for a miniature racecar using the PyBullet physics engine.
Other environments#
CompilerGym: Optimise compiler tasks#
Reinforcement learning environments for compiler optimization tasks, such as LLVM phase ordering, GCC flag tuning, and CUDA loop nest code generation.
CARL: context adaptive RL#
Configurable reinforcement learning environments for testing generalization, e.g. CartPole with variable pole lengths or Brax robots with different ground frictions.
gym-cellular-automata: Cellular Automata environments#
Environments where the agent interacts with Cellular Automata by changing its cell states.
gym-sokoban: 2D Transportation Puzzles#
The environment consists of transportation puzzles in which the player’s goal is to push all boxes to the warehouse’s storage locations.
DACBench: Benchmark Library for Dynamic Algorithm configuration#
Environments for hyperparameter configuration using RL. Includes cheap surrogate benchmarks as well as real-world algorithms.
NLPGym: A toolkit to develop RL agents to solve NLP tasks#
NLPGym provides interactive environments for standard NLP tasks such as sequence tagging, question answering, and sequence classification.
ShinRL: Environments for evaluating RL algorithms#
ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives (Deep RL Workshop 2021)
gymnax: Hardware Accelerated RL Environments#
RL Environments in JAX which allows for highly vectorised environments with support for a number of environments, Gym, MinAtari, bsuite and more.
gym-anytrading: Financial trading environments for FOREX and STOCKS#
AnyTrading is a collection of Gym environments for reinforcement learning-based trading algorithms with a great focus on simplicity, flexibility, and comprehensiveness.
gym-mtsim: Financial trading for MetaTrader 5 platform#
MtSim is a simulator for the MetaTrader 5 trading platform for reinforcement learning-based trading algorithms.
openmodelica-microgrid-gym: Environments for controlling power electronic converters in microgrids#
The OpenModelica Microgrid Gym (OMG) package is a software toolbox for the simulation and control optimization of microgrids based on energy conversion by power electronic converters.
mobile-env: Environments for coordination of wireless mobile networks#
An open, minimalist Gym environment for autonomous coordination in wireless mobile networks.
GymFC: A flight control tuning and training framework#
GymFC is a modular framework for synthesizing neuro-flight controllers. Has been used to generate policies for the world’s first open-source neural network flight control firmware Neuroflight.