498 research outputs found
Robust Place Recognition using an Imaging Lidar
We propose a methodology for robust, real-time place recognition using an
imaging lidar, which yields image-quality high-resolution 3D point clouds.
Utilizing the intensity readings of an imaging lidar, we project the point
cloud and obtain an intensity image. ORB feature descriptors are extracted from
the image and encoded into a bag-of-words vector. The vector, used to identify
the point cloud, is inserted into a database that is maintained by DBoW for
fast place recognition queries. The returned candidate is further validated by
matching visual feature descriptors. To reject matching outliers, we apply PnP,
which minimizes the reprojection error of visual features' positions in
Euclidean space with their correspondences in 2D image space, using RANSAC.
Combining the advantages from both camera and lidar-based place recognition
approaches, our method is truly rotation-invariant and can tackle reverse
revisiting and upside-down revisiting. The proposed method is evaluated on
datasets gathered from a variety of platforms over different scales and
environments. Our implementation is available at
https://git.io/imaging-lidar-place-recognitionComment: ICRA 202
PADLoC: LiDAR-Based Deep Loop Closure Detection and Registration using Panoptic Attention
A key component of graph-based SLAM systems is the ability to detect loop
closures in a trajectory to reduce the drift accumulated over time from the
odometry. Most LiDAR-based methods achieve this goal by using only the
geometric information, disregarding the semantics of the scene. In this work,
we introduce PADLoC, a LiDAR-based loop closure detection and registration
architecture comprising a shared 3D convolutional feature extraction backbone,
a global descriptor head for loop closure detection, and a novel
transformer-based head for point cloud matching and registration. We present
multiple methods for estimating the point-wise matching confidence based on
diversity indices. Additionally, to improve forward-backward consistency, we
propose the use of two shared matching and registration heads with their source
and target inputs swapped by exploiting that the estimated relative
transformations must be inverse of each other. Furthermore, we leverage
panoptic information during training in the form of a novel loss function that
reframes the matching problem as a classification task in the case of the
semantic labels and as a graph connectivity assignment for the instance labels.
We perform extensive evaluations of PADLoC on multiple real-world datasets
demonstrating that it achieves state-of-the-art performance. The code of our
work is publicly available at http://padloc.cs.uni-freiburg.de
LiDAR-Based Place Recognition For Autonomous Driving: A Survey
LiDAR-based place recognition (LPR) plays a pivotal role in autonomous
driving, which assists Simultaneous Localization and Mapping (SLAM) systems in
reducing accumulated errors and achieving reliable localization. However,
existing reviews predominantly concentrate on visual place recognition (VPR)
methods. Despite the recent remarkable progress in LPR, to the best of our
knowledge, there is no dedicated systematic review in this area. This paper
bridges the gap by providing a comprehensive review of place recognition
methods employing LiDAR sensors, thus facilitating and encouraging further
research. We commence by delving into the problem formulation of place
recognition, exploring existing challenges, and describing relations to
previous surveys. Subsequently, we conduct an in-depth review of related
research, which offers detailed classifications, strengths and weaknesses, and
architectures. Finally, we summarize existing datasets, commonly used
evaluation metrics, and comprehensive evaluation results from various methods
on public datasets. This paper can serve as a valuable tutorial for newcomers
entering the field of place recognition and for researchers interested in
long-term robot localization. We pledge to maintain an up-to-date project on
our website https://github.com/ShiPC-AI/LPR-Survey.Comment: 26 pages,13 figures, 5 table
Contour Context: Abstract Structural Distribution for 3D LiDAR Loop Detection and Metric Pose Estimation
This paper proposes \textit{Contour Context}, a simple, effective, and
efficient topological loop closure detection pipeline with accurate 3-DoF
metric pose estimation, targeting the urban utonomous driving scenario. We
interpret the Cartesian birds' eye view (BEV) image projected from 3D LiDAR
points as layered distribution of structures. To recover elevation information
from BEVs, we slice them at different heights, and connected pixels at each
level will form contours. Each contour is parameterized by abstract
information, e.g., pixel count, center position, covariance, and mean height.
The similarity of two BEVs is calculated in sequential discrete and continuous
steps. The first step considers the geometric consensus of graph-like
constellations formed by contours in particular localities. The second step
models the majority of contours as a 2.5D Gaussian mixture model, which is used
to calculate correlation and optimize relative transform in continuous space. A
retrieval key is designed to accelerate the search of a database indexed by
layered KD-trees. We validate the efficacy of our method by comparing it with
recent works on public datasets.Comment: 7 pages, 7 figures, accepted by ICRA 202
CVTNet: A Cross-View Transformer Network for Place Recognition Using LiDAR Data
LiDAR-based place recognition (LPR) is one of the most crucial components of
autonomous vehicles to identify previously visited places in GPS-denied
environments. Most existing LPR methods use mundane representations of the
input point cloud without considering different views, which may not fully
exploit the information from LiDAR sensors. In this paper, we propose a
cross-view transformer-based network, dubbed CVTNet, to fuse the range image
views (RIVs) and bird's eye views (BEVs) generated from the LiDAR data. It
extracts correlations within the views themselves using intra-transformers and
between the two different views using inter-transformers. Based on that, our
proposed CVTNet generates a yaw-angle-invariant global descriptor for each
laser scan end-to-end online and retrieves previously seen places by descriptor
matching between the current query scan and the pre-built database. We evaluate
our approach on three datasets collected with different sensor setups and
environmental conditions. The experimental results show that our method
outperforms the state-of-the-art LPR methods with strong robustness to
viewpoint changes and long-time spans. Furthermore, our approach has a good
real-time performance that can run faster than the typical LiDAR frame rate.
The implementation of our method is released as open source at:
https://github.com/BIT-MJY/CVTNet.Comment: accepted by IEEE Transactions on Industrial Informatics 202
A Survey on Global LiDAR Localization
Knowledge about the own pose is key for all mobile robot applications. Thus
pose estimation is part of the core functionalities of mobile robots. In the
last two decades, LiDAR scanners have become a standard sensor for robot
localization and mapping. This article surveys recent progress and advances in
LiDAR-based global localization. We start with the problem formulation and
explore the application scope. We then present the methodology review covering
various global localization topics, such as maps, descriptor extraction, and
consistency checks. The contents are organized under three themes. The first is
the combination of global place retrieval and local pose estimation. Then the
second theme is upgrading single-shot measurement to sequential ones for
sequential global localization. The third theme is extending single-robot
global localization to cross-robot localization on multi-robot systems. We end
this survey with a discussion of open challenges and promising directions on
global lidar localization
- …