364 research outputs found
Sparsity Invariant CNNs
In this paper, we consider convolutional neural networks operating on sparse
inputs with an application to depth upsampling from sparse laser scan data.
First, we show that traditional convolutional networks perform poorly when
applied to sparse data even when the location of missing data is provided to
the network. To overcome this problem, we propose a simple yet effective sparse
convolution layer which explicitly considers the location of missing data
during the convolution operation. We demonstrate the benefits of the proposed
network architecture in synthetic and real experiments with respect to various
baseline approaches. Compared to dense baselines, the proposed sparse
convolution network generalizes well to novel datasets and is invariant to the
level of sparsity in the data. For our evaluation, we derive a novel dataset
from the KITTI benchmark, comprising 93k depth annotated RGB images. Our
dataset allows for training and evaluating depth upsampling and depth
prediction techniques in challenging real-world settings and will be made
available upon publication
The Cityscapes Dataset for Semantic Urban Scene Understanding
Visual understanding of complex urban street scenes is an enabling factor for
a wide range of applications. Object detection has benefited enormously from
large-scale datasets, especially in the context of deep learning. For semantic
urban scene understanding, however, no current dataset adequately captures the
complexity of real-world urban scenes.
To address this, we introduce Cityscapes, a benchmark suite and large-scale
dataset to train and test approaches for pixel-level and instance-level
semantic labeling. Cityscapes is comprised of a large, diverse set of stereo
video sequences recorded in streets from 50 different cities. 5000 of these
images have high quality pixel-level annotations; 20000 additional images have
coarse annotations to enable methods that leverage large volumes of
weakly-labeled data. Crucially, our effort exceeds previous attempts in terms
of dataset size, annotation richness, scene variability, and complexity. Our
accompanying empirical study provides an in-depth analysis of the dataset
characteristics, as well as a performance evaluation of several
state-of-the-art approaches based on our benchmark.Comment: Includes supplemental materia
Code verification examples of a fully geometrical nonlinear membrane element using the method of manufactured solutions
This paper presents an effective method to perform Code Verification of a software which is designed for structural analysis using membranes. The focus lies on initially curved structures with large deformations in steady and unsteady regimes. The material is assumed to be linear elastic isotropic. Code Verification is a part of efforts to guarantee the code’s correctness and to obtain finally predictive capability of the code. The Method of Manufactured Solutions turned out to be an effective tool to perform Code Verification, especially for initially curved structures. Here arbitrary invented geometries and analytical solutions are chosen. The computer code must approach this solution asymptotically. The observed error reduction with systematic mesh refinement (i.e. observed order of accuracy) must be in the range of the formal order of accuracy, e.g. derived by a Taylor series expansion. If these two orders match in the asymptotic range, the implemented numerical algorithms are working as intended. The given examples provide a complete hierarchical benchmark suite for the reader to assess other codes, too. In the present case several membrane states were tested successfully and the used code Carat++ assessed to converge - as intended - second order accurately in space and time for all kind of shapes and solution
Cityscapes 3D: Dataset and Benchmark for 9 DoF Vehicle Detection
Detecting vehicles and representing their position and orientation in the
three dimensional space is a key technology for autonomous driving. Recently,
methods for 3D vehicle detection solely based on monocular RGB images gained
popularity. In order to facilitate this task as well as to compare and drive
state-of-the-art methods, several new datasets and benchmarks have been
published. Ground truth annotations of vehicles are usually obtained using
lidar point clouds, which often induces errors due to imperfect calibration or
synchronization between both sensors. To this end, we propose Cityscapes 3D,
extending the original Cityscapes dataset with 3D bounding box annotations for
all types of vehicles. In contrast to existing datasets, our 3D annotations
were labeled using stereo RGB images only and capture all nine degrees of
freedom. This leads to a pixel-accurate reprojection in the RGB image and a
higher range of annotations compared to lidar-based approaches. In order to
ease multitask learning, we provide a pairing of 2D instance segments with 3D
bounding boxes. In addition, we complement the Cityscapes benchmark suite with
3D vehicle detection based on the new annotations as well as metrics presented
in this work. Dataset and benchmark are available online.Comment: 2020 "Scalability in Autonomous Driving" CVPR Worksho
- …