295 research outputs found
RT-K-Net: Revisiting K-Net for Real-Time Panoptic Segmentation
Panoptic segmentation is one of the most challenging scene parsing tasks,
combining the tasks of semantic segmentation and instance segmentation. While
much progress has been made, few works focus on the real-time application of
panoptic segmentation methods. In this paper, we revisit the recently
introduced K-Net architecture. We propose vital changes to the architecture,
training, and inference procedure, which massively decrease latency and improve
performance. Our resulting RT-K-Net sets a new state-of-the-art performance for
real-time panoptic segmentation methods on the Cityscapes dataset and shows
promising results on the challenging Mapillary Vistas dataset. On Cityscapes,
RT-K-Net reaches 60.2 % PQ with an average inference time of 32 ms for full
resolution 1024x2048 pixel images on a single Titan RTX GPU. On Mapillary
Vistas, RT-K-Net reaches 33.2 % PQ with an average inference time of 69 ms.
Source code is available at https://github.com/markusschoen/RT-K-Net.Comment: Accepted at 2023 IEEE Intelligent Vehicles Symposiu
Extrinsic Infrastructure Calibration Using the Hand-Eye Robot-World Formulation
We propose a certifiably globally optimal approach for solving the hand-eye
robot-world problem supporting multiple sensors and targets at once. Further,
we leverage this formulation for estimating a geo-referenced calibration of
infrastructure sensors. Since vehicle motion recorded by infrastructure sensors
is mostly planar, obtaining a unique solution for the respective hand-eye
robot-world problem is unfeasible without incorporating additional knowledge.
Hence, we extend our proposed method to include a-priori knowledge, i.e., the
translation norm of calibration targets, to yield a unique solution. Our
approach achieves state-of-the-art results on simulated and real-world data.
Especially on real-world intersection data, our approach utilizing the
translation norm is the only method providing accurate results.Comment: Accepted at 2023 IEEE Intelligent Vehicles Symposiu
Identification of Threat Regions From a Dynamic Occupancy Grid Map for Situation-Aware Environment Perception
The advance towards higher levels of automation within the field of automated
driving is accompanied by increasing requirements for the operational safety of
vehicles. Induced by the limitation of computational resources, trade-offs
between the computational complexity of algorithms and their potential to
ensure safe operation of automated vehicles are often encountered.
Situation-aware environment perception presents one promising example, where
computational resources are distributed to regions within the perception area
that are relevant for the task of the automated vehicle. While prior map
knowledge is often leveraged to identify relevant regions, in this work, we
present a lightweight identification of safety-relevant regions that relies
solely on online information. We show that our approach enables safe vehicle
operation in critical scenarios, while retaining the benefits of non-uniformly
distributed resources within the environment perception.Comment: Accepted for publication at the 25th IEEE International Conference on
Intelligent Transportation Systems 2022. V2: added IEEE copyright notice V3:
Added DO
DeepCLR: Correspondence-Less Architecture for Deep End-to-End Point Cloud Registration
This work addresses the problem of point cloud registration using deep neural
networks. We propose an approach to predict the alignment between two point
clouds with overlapping data content, but displaced origins. Such point clouds
originate, for example, from consecutive measurements of a LiDAR mounted on a
moving platform. The main difficulty in deep registration of raw point clouds
is the fusion of template and source point cloud. Our proposed architecture
applies flow embedding to tackle this problem, which generates features that
describe the motion of each template point. These features are then used to
predict the alignment in an end-to-end fashion without extracting explicit
point correspondences between both input clouds. We rely on the KITTI odometry
and ModelNet40 datasets for evaluating our method on various point
distributions. Our approach achieves state-of-the-art accuracy and the lowest
run-time of the compared methods.Comment: 7 pages, 5 figures, 4 table
- …