SnapNet-R: Consistent 3D Multi-View Semantic Labeling for Robotics

Date:

Talk at ICCV’2017 3D Reconstruction meets Semantics workshop, located in Venice, Italy.

Abstract: We present a new approach for semantic recognition in the context of robotics. When a robot evolves in its environment, it gets 3D information given either by its sensors or by its own motion through 3D reconstruction. Our approach uses (i) 3D-coherent synthesis of scene observations and (ii) mix them in a multi-view framework for 3D labeling. (iii) This is efficient locally (for 2D semantic segmentation) and globally (for 3D structure labeling). This allows to add semantics to the observed scene that goes beyond simple image classification, as shown on challenging datasets such as SUNRGBD or the 3DRMS Reconstruction Challenge.

[ associated paper: cvf pdf #1 ]