Skip to content

Commit 3d5c218

Browse files
authored
Update third party envs (#404)
1 parent e619e1b commit 3d5c218

File tree

1 file changed

+43
-37
lines changed

1 file changed

+43
-37
lines changed

Diff for: docs/environments/third_party_environments.md

+43-37
Original file line numberDiff line numberDiff line change
@@ -45,6 +45,13 @@ PyBullet based simulations of a robotic arm moving objects.
4545

4646
tmrl is a distributed framework for training Deep Reinforcement Learning AIs in real-time applications. It is demonstrated on the TrackMania 2020 video game.
4747

48+
### [gym-jiminy: Training Robots in Jiminy](https://github.com/duburcqa/jiminy)
49+
50+
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.28.0-blue)]()
51+
[![GitHub stars](https://img.shields.io/github/stars/duburcqa/jiminy)]()
52+
53+
gym-jiminy presents an extension of the initial Gym for robotics using [Jiminy](https://github.com/duburcqa/jiminy), an extremely fast and light-weight simulator for poly-articulated systems using Pinocchio for physics evaluation and Meshcat for web-based 3D rendering.
54+
4855
### [Safety-Gymnasium: Ensuring safety in real-world RL scenarios](https://github.com/PKU-MARL/safety-gymnasium)
4956

5057
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.3-blue)]()
@@ -57,7 +64,7 @@ Highly scalable and customizable Safe Reinforcement Learning library.
5764
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
5865
[![GitHub stars](https://img.shields.io/github/stars/MatPoliquin/stable-retro)]()
5966

60-
Supported fork of gym-retro: turn classic video games into Gymnasium environments.
67+
Supported fork of [gym-retro](https://openai.com/research/gym-retro): turn classic video games into Gymnasium environments.
6168

6269
### [flappy-bird-gymnasium: A Flappy Bird environment for Gymnasium](https://github.com/markub3327/flappy-bird-gymnasium)
6370

@@ -66,6 +73,13 @@ Supported fork of gym-retro: turn classic video games into Gymnasium environment
6673

6774
A simple environment for single-agent reinforcement learning algorithms on a clone of [Flappy Bird](https://en.wikipedia.org/wiki/Flappy_Bird), the hugely popular arcade-style mobile game. Both state and pixel observation environments are available.
6875

76+
### [gym-saturation: Environments used to prove theorems](https://github.com/inpefess/gym-saturation)
77+
78+
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
79+
[![GitHub stars](https://img.shields.io/github/stars/inpefess/gym-saturation)]()
80+
81+
An environment for guiding automated theorem provers based on saturation algorithms (e.g. [Vampire](https://github.com/vprover/vampire)).
82+
6983
### [matrix-mdp: Easily create discrete MDPs](https://github.com/Paul-543NA/matrix-mdp-gym)
7084

7185
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.2-blue)]()
@@ -82,26 +96,22 @@ Many of these can be adapted to work with gymnasium (see [Compatibility with Gym
8296

8397
### [gym-derk: GPU accelerated MOBA environment](https://gym.derkgame.com/)
8498

85-
This is a 3v3 MOBA environment where you train creatures to fight each other. It runs entirely on the GPU so you can easily have hundreds of instances running in parallel. There are around 15 items for the creatures, 60 "senses", 5 actions, and roughly 23 tweakable rewards. It's also possible to benchmark an agent against other agents online. It's available for free for training for personal use, and otherwise costs money; see licensing details on the website
99+
A 3v3 MOBA environment where you train creatures to fight each other.
86100

87-
### [SlimeVolleyGym: A simple environment for single and multi-agent reinforcement learning](https://github.com/hardmaru/slimevolleygym)
101+
### [SlimeVolleyGym: A simple environment for Slime Volleyball game](https://github.com/hardmaru/slimevolleygym)
88102

89-
A simple environment for benchmarking single and multi-agent reinforcement learning algorithms on a clone of Slime Volleyball game. The only dependencies are gym and NumPy. Both state and pixel observation environments are available. The motivation of this environment is to easily enable trained agents to play against each other, and also facilitate the training of agents directly in a multi-agent setting, thus adding an extra dimension for evaluating an agent's performance.
103+
A simple environment for benchmarking single and multi-agent reinforcement learning algorithms on a clone of Slime Volleyball game.
90104

91105
### [Unity ML Agents: Environments for Unity game engine](https://github.com/Unity-Technologies/ml-agents)
92106

93-
Gym wrappers for arbitrary and premade environments with the Unity game engine.
107+
Gym (and PettingZoo) wrappers for arbitrary and premade environments with the Unity game engine.
94108

95109
### [PGE: Parallel Game Engine](https://github.com/222464/PGE)
96110

97-
PGE is a FOSS 3D engine for AI simulations and can interoperate with the Gym. Contains environments with modern 3D graphics, and uses Bullet for physics.
111+
Uses The [Open 3D Engine](https://www.o3de.org/) for AI simulations and can interoperate with the Gym. Uses [PyBullet](https://github.com/bulletphysics/bullet3) physics.
98112

99113
## Robotics environments
100114

101-
### [gym-jiminy: Training Robots in Jiminy](https://github.com/duburcqa/jiminy)
102-
103-
gym-jiminy presents an extension of the initial Gym for robotics using Jiminy, an extremely fast and light-weight simulator for poly-articulated systems using Pinocchio for physics evaluation and Meshcat for web-based 3D rendering.
104-
105115
### [gym-pybullet-drones: Environments for quadcopter control](https://github.com/JacopoPan/gym-pybullet-drones)
106116

107117
A simple environment using [PyBullet](https://github.com/bulletphysics/bullet3) to simulate the dynamics of a [Bitcraze Crazyflie 2.x](https://www.bitcraze.io/documentation/hardware/crazyflie_2_1/crazyflie_2_1-datasheet.pdf) nanoquadrotor.
@@ -115,55 +125,55 @@ Mars Explorer is a Gym compatible environment designed and developed as an initi
115125

116126
Robo-gym provides a collection of reinforcement learning environments involving robotic tasks applicable in both simulation and real-world robotics.
117127

118-
### [Offworld-gym](https://github.com/offworld-projects/offworld-gym)
128+
### [Offworld-gym: Control real robots remotely for free](https://github.com/offworld-projects/offworld-gym)
119129

120-
Gym environments that let you control physics robotics in a laboratory via the internet.
130+
Gym environments that let you control real robots in a laboratory via the internet.
121131

122-
### [safe-control-gym](https://github.com/utiasDSL/safe-control-gym)
132+
### [safe-control-gym: Evaluate safety of RL algorithms](https://github.com/utiasDSL/safe-control-gym)
123133

124-
PyBullet based CartPole and Quadrotor environments—with [CasADi](https://web.casadi.org) (symbolic) *a priori* dynamics and constraints—for learning-based control and model-based reinforcement learning.
134+
Evaluate safety, robustness and generalization via PyBullet based CartPole and Quadrotor environments—with [CasADi](https://web.casadi.org) (symbolic) *a priori* dynamics and constraints.
125135

126136
### [gym-softrobot: Soft-robotics environments](https://github.com/skim0119/gym-softrobot/)
127137

128138
A large-scale benchmark for co-optimizing the design and control of soft robots.
129139

130-
### [iGibson](https://github.com/StanfordVL/iGibson/)
140+
### [iGibson: Photorealistic and interactive robotics environments](https://github.com/StanfordVL/iGibson/)
131141

132-
A Simulation Environment to train Robots in Large Realistic Interactive Scenes
142+
A simulation environment with high-quality realistic scenes, with interactive physics using [PyBullet](https://github.com/bulletphysics/bullet3).
133143

134-
### [DexterousHands: dual dexterous hand manipulation tasks](https://github.com/PKU-MARL/DexterousHands/)
144+
### [DexterousHands: Dual dexterous hand manipulation tasks](https://github.com/PKU-MARL/DexterousHands/)
135145

136-
This is a library that provides dual dexterous hand manipulation tasks through Isaac Gym
146+
This is a library that provides dual dexterous hand manipulation tasks through Isaac Gym.
137147

138-
### [OmniIsaacGymEnvs](https://github.com/NVIDIA-Omniverse/OmniIsaacGymEnvs/)
148+
### [OmniIsaacGymEnvs: Gym environments for NVIDIA Omniverse Isaac ](https://github.com/NVIDIA-Omniverse/OmniIsaacGymEnvs/)
139149

140-
Reinforcement Learning Environments for Omniverse Isaac Gym
150+
Reinforcement Learning Environments for [Omniverse Isaac simulator](https://docs.omniverse.nvidia.com/app_isaacsim/app_isaacsim/overview.html).
141151

142152
## Autonomous Driving environments
143153

144-
### [gym-duckietown](https://github.com/duckietown/gym-duckietown)
154+
### [gym-duckietown: Lane-following for self-driving cars](https://github.com/duckietown/gym-duckietown)
145155

146156
A lane-following simulator built for the [Duckietown](http://duckietown.org/) project (small-scale self-driving car course).
147157

148-
### [gym-electric-motor](https://github.com/upb-lea/gym-electric-motor)
158+
### [gym-electric-motor: Gym environments for electric motor simulations](https://github.com/upb-lea/gym-electric-motor)
149159

150-
An environment for simulating a wide variety of electric drives taking into account different types of electric motors and converters. Control schemes can be continuous, yielding a voltage duty cycle, or discrete, determining converter switching states directly.
160+
An environment for simulating a wide variety of electric drives taking into account different types of electric motors and converters.
151161

152-
### [CommonRoad-RL](https://commonroad.in.tum.de/tools/commonroad-rl)
162+
### [CommonRoad-RL: Motion planning for traffic scenarios ](https://commonroad.in.tum.de/tools/commonroad-rl)
153163

154164
A Gym for solving motion planning problems for various traffic scenarios compatible with [CommonRoad benchmarks](https://commonroad.in.tum.de/scenarios), which provides configurable rewards, action spaces, and observation spaces.
155165

156-
### [racing_dreamer](https://github.com/CPS-TUWien/racing_dreamer/)
166+
### [racing_dreamer: Latent imagination in autonomous racing](https://github.com/CPS-TUWien/racing_dreamer/)
157167

158-
Latent Imagination Facilitates Zero-Shot Transfer in Autonomous Racing
168+
Train a model-based RL agent in simulation and, without finetuning, transfer it to small-scale race cars.
159169

160170
### [l2r: Multimodal control environment where agents learn how to race](https://github.com/learn-to-race/l2r/)
161171

162172
An open-source reinforcement learning environment for autonomous racing.
163173

164-
### [racecar_gym](https://github.com/axelbr/racecar_gym/)
174+
### [racecar_gym: Miniature racecar env using PyBullet](https://github.com/axelbr/racecar_gym/)
165175

166-
A gym environment for a miniature racecar using the pybullet physics engine.
176+
A gym environment for a miniature racecar using the [PyBullet](https://github.com/bulletphysics/bullet3) physics engine.
167177

168178
## Other environments
169179

@@ -181,19 +191,15 @@ Environments where the agent interacts with _Cellular Automata_ by changing its
181191

182192
### [gym-sokoban: 2D Transportation Puzzles](https://github.com/mpSchrader/gym-sokoban)
183193

184-
The environment consists of transportation puzzles in which the player's goal is to push all boxes to the warehouse's storage locations. The advantage of the environment is that it generates a new random level every time it is initialized or reset, which prevents overfitting to predefined levels.
194+
The environment consists of transportation puzzles in which the player's goal is to push all boxes to the warehouse's storage locations.
185195

186196
### [DACBench: Benchmark Library for Dynamic Algorithm configuration](https://github.com/automl/DACBench)
187197

188-
Environments for hyperparameter configuration using RL. Includes cheap surrogate benchmarks as well as real-world algorithms from e.g. AI Planning, Evolutionary Computation and Deep Learning.
198+
Environments for hyperparameter configuration using RL. Includes cheap surrogate benchmarks as well as real-world algorithms.
189199

190200
### [NLPGym: A toolkit to develop RL agents to solve NLP tasks](https://github.com/rajcscw/nlp-gym)
191201

192-
[NLPGym](https://arxiv.org/pdf/2011.08272v1.pdf) provides interactive environments for standard NLP tasks such as sequence tagging, question answering, and sequence classification. Users can easily customize the tasks with their datasets, observations, features and reward functions.
193-
194-
### [gym-saturation: Environments used to prove theorems](https://github.com/inpefess/gym-saturation)
195-
196-
An environment for guiding automated theorem provers based on saturation algorithms (e.g. [Vampire](https://github.com/vprover/vampire)).
202+
[NLPGym](https://arxiv.org/pdf/2011.08272v1.pdf) provides interactive environments for standard NLP tasks such as sequence tagging, question answering, and sequence classification.
197203

198204
### [ShinRL: Environments for evaluating RL algorithms](https://github.com/omron-sinicx/ShinRL/)
199205

@@ -209,7 +215,7 @@ AnyTrading is a collection of Gym environments for reinforcement learning-based
209215

210216
### [gym-mtsim: Financial trading for MetaTrader 5 platform](https://github.com/AminHP/gym-mtsim)
211217

212-
MtSim is a simulator for the MetaTrader 5 trading platform for reinforcement learning-based trading algorithms. MetaTrader 5 is a multi-asset platform that allows trading Forex, Stocks, Crypto, and Futures.
218+
MtSim is a simulator for the [MetaTrader 5](https://www.metatrader5.com/) trading platform for reinforcement learning-based trading algorithms.
213219

214220
### [openmodelica-microgrid-gym: Environments for controlling power electronic converters in microgrids](https://github.com/upb-lea/openmodelica-microgrid-gym)
215221

@@ -221,4 +227,4 @@ An open, minimalist Gym environment for autonomous coordination in wireless mobi
221227

222228
### [GymFC: A flight control tuning and training framework](https://github.com/wil3/gymfc/)
223229

224-
GymFC is a modular framework for synthesizing neuro-flight controllers. The architecture integrates digital twinning concepts to provide a seamless transfer of trained policies to hardware. The environment has been used to generate policies for the world's first open-source neural network flight control firmware [Neuroflight](https://github.com/wil3/neuroflight).
230+
GymFC is a modular framework for synthesizing neuro-flight controllers. Has been used to generate policies for the world's first open-source neural network flight control firmware [Neuroflight](https://github.com/wil3/neuroflight).

0 commit comments

Comments
 (0)