Visual And Tactile 3d Point Cloud Data From Real Robots For Shape Modeling And Completion

DATA IN BRIEF(2020)

引用 2|浏览94
暂无评分
摘要
Representing 3D geometry for different tasks, e.g. rendering and reconstruction, is an important goal in different fields, such as computer graphics, computer vision and robotics. Robotic applications often require perception of object shape information extracted from sensory data that can be noisy and incomplete. This is a challenging task and in order to facilitate analysis of new methods and comparison of different approaches for shape modeling (e.g. surface estimation), completion and exploration, we provide real sensory data acquired from exploring various objects of different complexities. The dataset includes visual and tactile readings in the form of 3D point clouds obtained using two different robot setups that are equipped with visual and tactile sensors. During data collection, the robots touch the experiment objects in a predefined manner at various exploration configurations and gather visual and tactile points in the same coordinate frame based on calibration between the robots and the used cameras. The goal of this exhaustive exploration procedure is to sense unseen parts of the objects which are not visible to the cameras, but can be sensed via tactile sensors activated at touched areas. The data was used for shape completion and modeling via Implicit Surface representation and Gaussian-Process-based regression, in the work "Object shape estimation and modeling, based on sparse Gaussian process implicit surfaces, combining visual data and tactile exploration" [3], and also used partially in "Enhancing visual perception of shape through tactile glances" [4], both studying efficient exploration of objects to reduce number of touches. (C) 2020 Published by Elsevier Inc.
更多
查看译文
关键词
Point cloud, Tactile sensing, Visual depth sensing, Shape modeling, Shape completion, Implicit surface, Gaussian process, Regression
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要