Visual and tactile 3D point cloud data from real robots for shape modeling and completion.
Ontology highlight
ABSTRACT: Representing 3D geometry for different tasks, e.g. rendering and reconstruction, is an important goal in different fields, such as computer graphics, computer vision and robotics. Robotic applications often require perception of object shape information extracted from sensory data that can be noisy and incomplete. This is a challenging task and in order to facilitate analysis of new methods and comparison of different approaches for shape modeling (e.g. surface estimation), completion and exploration, we provide real sensory data acquired from exploring various objects of different complexities. The dataset includes visual and tactile readings in the form of 3D point clouds obtained using two different robot setups that are equipped with visual and tactile sensors. During data collection, the robots touch the experiment objects in a predefined manner at various exploration configurations and gather visual and tactile points in the same coordinate frame based on calibration between the robots and the used cameras. The goal of this exhaustive exploration procedure is to sense unseen parts of the objects which are not visible to the cameras, but can be sensed via tactile sensors activated at touched areas. The data was used for shape completion and modeling via Implicit Surface representation and Gaussian-Process-based regression, in the work "Object shape estimation and modeling, based on sparse Gaussian process implicit surfaces, combining visual data and tactile exploration" [3], and also used partially in "Enhancing visual perception of shape through tactile glances" [4], both studying efficient exploration of objects to reduce number of touches.
SUBMITTER: Bekiroglu Y
PROVIDER: S-EPMC7125316 | biostudies-literature | 2020 Jun
REPOSITORIES: biostudies-literature
ACCESS DATA