2,234 research outputs found
Matterport3D: Learning from RGB-D Data in Indoor Environments
Access to large, diverse RGB-D datasets is critical for training RGB-D scene
understanding algorithms. However, existing datasets still cover only a limited
number of views or a restricted scale of spaces. In this paper, we introduce
Matterport3D, a large-scale RGB-D dataset containing 10,800 panoramic views
from 194,400 RGB-D images of 90 building-scale scenes. Annotations are provided
with surface reconstructions, camera poses, and 2D and 3D semantic
segmentations. The precise global alignment and comprehensive, diverse
panoramic set of views over entire buildings enable a variety of supervised and
self-supervised computer vision tasks, including keypoint matching, view
overlap prediction, normal prediction from color, semantic segmentation, and
region classification
Multi-View Deep Learning for Consistent Semantic Mapping with RGB-D Cameras
Visual scene understanding is an important capability that enables robots to
purposefully act in their environment. In this paper, we propose a novel
approach to object-class segmentation from multiple RGB-D views using deep
learning. We train a deep neural network to predict object-class semantics that
is consistent from several view points in a semi-supervised way. At test time,
the semantics predictions of our network can be fused more consistently in
semantic keyframe maps than predictions of a network trained on individual
views. We base our network architecture on a recent single-view deep learning
approach to RGB and depth fusion for semantic object-class segmentation and
enhance it with multi-scale loss minimization. We obtain the camera trajectory
using RGB-D SLAM and warp the predictions of RGB-D images into ground-truth
annotated frames in order to enforce multi-view consistency during training. At
test time, predictions from multiple views are fused into keyframes. We propose
and analyze several methods for enforcing multi-view consistency during
training and testing. We evaluate the benefit of multi-view consistency
training and demonstrate that pooling of deep features and fusion over multiple
views outperforms single-view baselines on the NYUDv2 benchmark for semantic
segmentation. Our end-to-end trained network achieves state-of-the-art
performance on the NYUDv2 dataset in single-view segmentation as well as
multi-view semantic fusion.Comment: the 2017 IEEE/RSJ International Conference on Intelligent Robots and
Systems (IROS 2017
SEGCloud: Semantic Segmentation of 3D Point Clouds
3D semantic scene labeling is fundamental to agents operating in the real
world. In particular, labeling raw 3D point sets from sensors provides
fine-grained semantics. Recent works leverage the capabilities of Neural
Networks (NNs), but are limited to coarse voxel predictions and do not
explicitly enforce global consistency. We present SEGCloud, an end-to-end
framework to obtain 3D point-level segmentation that combines the advantages of
NNs, trilinear interpolation(TI) and fully connected Conditional Random Fields
(FC-CRF). Coarse voxel predictions from a 3D Fully Convolutional NN are
transferred back to the raw 3D points via trilinear interpolation. Then the
FC-CRF enforces global consistency and provides fine-grained semantics on the
points. We implement the latter as a differentiable Recurrent NN to allow joint
optimization. We evaluate the framework on two indoor and two outdoor 3D
datasets (NYU V2, S3DIS, KITTI, Semantic3D.net), and show performance
comparable or superior to the state-of-the-art on all datasets.Comment: Accepted as a spotlight at the International Conference of 3D Vision
(3DV 2017
3D-BEVIS: Bird's-Eye-View Instance Segmentation
Recent deep learning models achieve impressive results on 3D scene analysis
tasks by operating directly on unstructured point clouds. A lot of progress was
made in the field of object classification and semantic segmentation. However,
the task of instance segmentation is less explored. In this work, we present
3D-BEVIS, a deep learning framework for 3D semantic instance segmentation on
point clouds. Following the idea of previous proposal-free instance
segmentation approaches, our model learns a feature embedding and groups the
obtained feature space into semantic instances. Current point-based methods
scale linearly with the number of points by processing local sub-parts of a
scene individually. However, to perform instance segmentation by clustering,
globally consistent features are required. Therefore, we propose to combine
local point geometry with global context information from an intermediate
bird's-eye view representation.Comment: camera-ready version for GCPR '1
Exploiting 2D Floorplan for Building-scale Panorama RGBD Alignment
This paper presents a novel algorithm that utilizes a 2D floorplan to align
panorama RGBD scans. While effective panorama RGBD alignment techniques exist,
such a system requires extremely dense RGBD image sampling. Our approach can
significantly reduce the number of necessary scans with the aid of a floorplan
image. We formulate a novel Markov Random Field inference problem as a scan
placement over the floorplan, as opposed to the conventional scan-to-scan
alignment. The technical contributions lie in multi-modal image correspondence
cues (between scans and schematic floorplan) as well as a novel coverage
potential avoiding an inherent stacking bias. The proposed approach has been
evaluated on five challenging large indoor spaces. To the best of our
knowledge, we present the first effective system that utilizes a 2D floorplan
image for building-scale 3D pointcloud alignment. The source code and the data
will be shared with the community to further enhance indoor mapping research
- …