You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently, doing Reinforcement Learning in Gazebo is really hard. It is not impossible (as we already have reset methods), but as RL gains traction in the Robotics Community we are seeing real world use cases where walking models are being developed on other simulators. It would be nice to provide users with a way of doing RL. In particular, perhaps an example which exposes a gym api for one of the major RL libraries (stable baselines, rllib, etc). This work would also necessitate some improvements to the python API and some work on the python packaging side. In particular my proposal is we do the following:
Desired behavior
Currently, doing Reinforcement Learning in Gazebo is really hard. It is not impossible (as we already have reset methods), but as RL gains traction in the Robotics Community we are seeing real world use cases where walking models are being developed on other simulators. It would be nice to provide users with a way of doing RL. In particular, perhaps an example which exposes a gym api for one of the major RL libraries (stable baselines, rllib, etc). This work would also necessitate some improvements to the python API and some work on the python packaging side. In particular my proposal is we do the following:
ISystemReset
in test fixture #2647OR
PostUpdate
,Update
andPreUpdate
.Stretch goals
Some relevant projects
Few Other Design Considerations
The work done on the python bindings actually exposes a really nice API. It is also nice that we have a separation between client and server API.
The text was updated successfully, but these errors were encountered: