Terrible name, but very interesting dataset:
GitHub describes the input quite well:
The model takes as input a RGB image from the robot workspace camera and a task string describing the task that the robot is supposed to perform.What task the model should perform is communicated to the model purely through the task string. The image communicates to the model the current state of the world, i.e. assuming the model runs at three hertz, every 333 milliseconds, we feed the latest RGB image from a robot workspace camera into the model to obtain the next action to take.
TODO: how is the scenario specified?
TODO: any simulation integration to it?
Python library for Bullet Physics.
Became very popular as of result of people using Bullet Physics for reinforcement learning AI training robot simulations.
Website: pybullet.org/
Source code: somewhere inside the main Bullet Physics source tree. Yay.