12 research outputs found
Handling Constrained Optimization in Factor Graphs for Autonomous Navigation
Factor graphs are graphical models used to represent
a wide variety of problems across robotics, such as Structure from
Motion (SfM), Simultaneous Localization and Mapping (SLAM)
and calibration. Typically, at their core, they have an optimization
problem whose terms only depend on a small subset of variables.
Factor graph solvers exploit the locality of problems to drastically
reduce the computational time of the Iterative Least-Squares (ILS)
methodology. Although extremely powerful, their application is
usually limited to unconstrained problems. In this letter, we model
constraints over variables within factor graphs by introducing a
factor graph version of the Augmented Lagrangian (AL) method.
We show the potential of our method by presenting a full navigation
stack based on factor graphs. Differently from standard navigation
stacks, we can model both optimal control for local planning and localization with factor graphs, and solve the two problems using the
standard ILS methodology.We validate our approach in real-world
autonomous navigation scenarios, comparing it with the de facto
standard navigation stack implemented in ROS. Comparative experiments show that for the application at hand our system outperforms the standard nonlinear programming solver Interior-Point
Optimizer (IPOPT) in runtime, while achieving similar solutions
Long-Term Localization using Semantic Cues in Floor Plan Maps
Lifelong localization in a given map is an essential capability for
autonomous service robots. In this paper, we consider the task of long-term
localization in a changing indoor environment given sparse CAD floor plans. The
commonly used pre-built maps from the robot sensors may increase the cost and
time of deployment. Furthermore, their detailed nature requires that they are
updated when significant changes occur. We address the difficulty of
localization when the correspondence between the map and the observations is
low due to the sparsity of the CAD map and the changing environment. To
overcome both challenges, we propose to exploit semantic cues that are commonly
present in human-oriented spaces. These semantic cues can be detected using RGB
cameras by utilizing object detection, and are matched against an
easy-to-update, abstract semantic map. The semantic information is integrated
into a Monte Carlo localization framework using a particle filter that operates
on 2D LiDAR scans and camera data. We provide a long-term localization solution
and a semantic map format, for environments that undergo changes to their
interior structure and detailed geometric maps are not available. We evaluate
our localization framework on multiple challenging indoor scenarios in an
office environment, taken weeks apart. The experiments suggest that our
approach is robust to structural changes and can run on an onboard computer. We
released the open source implementation of our approach written in C++ together
with a ROS wrapper.Comment: Under review for RA-
MD-SLAM: Multi-cue Direct SLAM
Simultaneous Localization and Mapping (SLAM) systems are fundamental building blocks for any autonomous robot navigating in unknown environments. The SLAM implementation heavily depends on the sensor modality employed on the mobile platform. For this reason, assumptions on the scene's structure are often made to maximize estimation accuracy. This paper presents a novel direct 3D SLAM pipeline that works independently for RGB-D and LiDAR sensors. Building upon prior work on multi-cue photometric frame-to-frame alignment [4], our proposed approach provides an easy-to-extend and generic SLAM system. Our pipeline requires only minor adaptations within the projection model to handle different sensor modalities. We couple a position tracking system with an appearance-based relocalization mechanism that handles large loop closures. Loop closures are validated by the same direct registration algorithm used for odometry estimation. We present comparative experiments with state-of-the-art approaches on publicly available benchmarks using RGB-D cameras and 3D LiDARs. Our system performs well in heterogeneous datasets compared to other sensor-specific methods while making no assumptions about the environment. Finally, we release an open-source C++ implementation of our system
LIO-EKF: High Frequency LiDAR-Inertial Odometry using Extended Kalman Filters
Odometry estimation is crucial for every autonomous system requiring
navigation in an unknown environment. In modern mobile robots, 3D
LiDAR-inertial systems are often used for this task. By fusing LiDAR scans and
IMU measurements, these systems can reduce the accumulated drift caused by
sequentially registering individual LiDAR scans and provide a robust pose
estimate. Although effective, LiDAR-inertial odometry systems require proper
parameter tuning to be deployed. In this paper, we propose LIO-EKF, a
tightly-coupled LiDAR-inertial odometry system based on point-to-point
registration and the classical extended Kalman filter scheme. We propose an
adaptive data association that considers the relative pose uncertainty, the map
discretization errors, and the LiDAR noise. In this way, we can substantially
reduce the parameters to tune for a given type of environment. The experimental
evaluation suggests that the proposed system performs on par with the
state-of-the-art LiDAR-inertial odometry pipelines but is significantly faster
in computing the odometry. The source code of our implementation is publicly
available (https://github.com/YibinWu/LIO-EKF).Comment: 7 pages, 2 figure
KISS-ICP: In Defense of Point-to-Point ICP -- Simple, Accurate, and Robust Registration If Done the Right Way
Robust and accurate pose estimation of a robotic platform, so-called
sensor-based odometry, is an essential part of many robotic applications. While
many sensor odometry systems made progress by adding more complexity to the
ego-motion estimation process, we move in the opposite direction. By removing a
majority of parts and focusing on the core elements, we obtain a surprisingly
effective system that is simple to realize and can operate under various
environmental conditions using different LiDAR sensors. Our odometry estimation
approach relies on point-to-point ICP combined with adaptive thresholding for
correspondence matching, a robust kernel, a simple but widely applicable motion
compensation approach, and a point cloud subsampling strategy. This yields a
system with only a few parameters that in most cases do not even have to be
tuned to a specific LiDAR sensor. Our system using the same parameters performs
on par with state-of-the-art methods under various operating conditions using
different platforms: automotive platforms, UAV-based operation, vehicles like
segways, or handheld LiDARs. We do not require integrating IMU information and
solely rely on 3D point cloud data obtained from a wide range of 3D LiDAR
sensors, thus, enabling a broad spectrum of different applications and
operating conditions. Our open-source system operates faster than the sensor
frame rate in all presented datasets and is designed for real-world scenarios.Comment: 8 page
Plug-and-Play SLAM: A Unified SLAM Architecture for Modularity and Ease of Use
Nowadays, SLAM (Simultaneous Localization and Mapping) is considered by the
Robotics community to be a mature field. Currently, there are many open-source
systems that are able to deliver fast and accurate estimation in typical
real-world scenarios. Still, all these systems often provide an ad-hoc
implementation that entailed to predefined sensor configurations. In this work,
we tackle this issue, proposing a novel SLAM architecture specifically designed
to address heterogeneous sensors' configuration and to standardize SLAM
solutions. Thanks to its modularity and to specific design patterns, the
presented architecture is easy to extend, enhancing code reuse and efficiency.
Finally, adopting our solution, we conducted comparative experiments for a
variety of sensor configurations, showing competitive results that confirm
state-of-the-art performance
Tree instance segmentation and traits estimation for forestry environments exploiting LiDAR data collected by mobile robots
Forests play a crucial role in our ecosystems, functioning as carbon sinks, climate stabilizers, biodiversity hubs, and sources of wood. By the very nature of their scale, monitoring and maintaining forests is a challenging task. Robotics in forestry can have the potential for substantial automation toward efficient and sustainable foresting practices. In this paper, we address the problem of automatically producing a forest inventory by exploiting LiDAR data collected by a mobile platform. To construct an inventory, we first extract tree instances from point clouds. Then, we process each instance to extract forestry inventory information. Our approach provides the per-tree geometric trait of “diameter at breast height” together with the individual tree locations in a plot. We validate our results against manual measurements collected by foresters during field trials. Our experiments show strong segmentation and tree trait estimation performance, underlining the potential for automating forestry services. Results furthermore show a superior performance compared to the popular baseline methods used in this domain
VDBFusion: Flexible and Efficient TSDF Integration of Range Sensor Data
Mapping is a crucial task in robotics and a fundamental building block of most mobile systems deployed in the real world. Robots use different environment representations depending on their task and sensor setup. This paper showcases a practical approach to volumetric surface reconstruction based on truncated signed distance functions, also called TSDFs. We revisit the basics of this mapping technique and offer an approach for building effective and efficient real-world mapping systems. In contrast to most state-of-the-art SLAM and mapping approaches, we are making no assumptions on the size of the environment nor the employed range sensor. Unlike most other approaches, we introduce an effective system that works in multiple domains using different sensors. To achieve this, we build upon the Academy-Award-winning OpenVDB library used in filmmaking to realize an effective 3D map representation. Based on this, our proposed system is flexible and highly effective and, in the end, capable of integrating point clouds from a 64-beam LiDAR sensor at 20 frames per second using a single-core CPU. Along with this publication comes an easy-to-use C++ and Python library to quickly and efficiently solve volumetric mapping problems with TSDFs
Robust Onboard Localization in Changing Environments Exploiting Text Spotting
Robust localization in a given map is a crucial component of most autonomous
robots. In this paper, we address the problem of localizing in an indoor
environment that changes and where prominent structures have no correspondence
in the map built at a different point in time. To overcome the discrepancy
between the map and the observed environment caused by such changes, we exploit
human-readable localization cues to assist localization. These cues are readily
available in most facilities and can be detected using RGB camera images by
utilizing text spotting. We integrate these cues into a Monte Carlo
localization framework using a particle filter that operates on 2D LiDAR scans
and camera data. By this, we provide a robust localization solution for
environments with structural changes and dynamics by humans walking. We
evaluate our localization framework on multiple challenging indoor scenarios in
an office environment. The experiments suggest that our approach is robust to
structural changes and can run on an onboard computer. We release an open
source implementation of our approach (upon paper acceptance), which uses
off-the-shelf text spotting, written in C++ with a ROS wrapper.Comment: This work has been accepted to IROS 2022. Copyright may be
transferred without notice, after which this version may no longer be
accessibl