75 research outputs found
Long-Term Urban Vehicle Localization Using Pole Landmarks Extracted from 3-D Lidar Scans
Due to their ubiquity and long-term stability, pole-like objects are well
suited to serve as landmarks for vehicle localization in urban environments. In
this work, we present a complete mapping and long-term localization system
based on pole landmarks extracted from 3-D lidar data. Our approach features a
novel pole detector, a mapping module, and an online localization module, each
of which are described in detail, and for which we provide an open-source
implementation at www.github.com/acschaefer/polex. In extensive experiments, we
demonstrate that our method improves on the state of the art with respect to
long-term reliability and accuracy: First, we prove reliability by tasking the
system with localizing a mobile robot over the course of 15~months in an urban
area based on an initial map, confronting it with constantly varying routes,
differing weather conditions, seasonal changes, and construction sites. Second,
we show that the proposed approach clearly outperforms a recently published
method in terms of accuracy.Comment: 9 page
Orientation-Aware 3D SLAM in Alternating Magnetic Field from Powerlines
Identifying new sensing modalities for indoor localization is an interest of research. This paper studies powerline-induced alternating magnetic field (AMF) that fills the indoor space for the orientation-aware three-dimensional (3D) simultaneous localization and mapping (SLAM). While an existing study has adopted a uniaxial AMF sensor for SLAM in a plane surface, the design falls short of addressing the vector field nature of AMF and is therefore susceptible to sensor orientation variations. Moreover, although the higher spatial variability of AMF in comparison with indoor geomagnetism promotes location sensing resolution, extra SLAM algorithm designs are needed to achieve robustness to trajectory deviations from the constructed map. To address the above issues, we design a new triaxial AMF sensor and a new SLAM algorithm that constructs a 3D AMF intensity map regularized and augmented by a Gaussian process. The triaxial sensor’s orientation estimation is free of the error accumulation problem faced by inertial sensing. From extensive evaluation in eight indoor environments, our AMF-based 3D SLAM achieves sub-1m to 3m median localization errors in spaces of up to 500 m2 , sub-2° mean error in orientation sensing, and outperforms the SLAM systems based on Wi-Fi, geomagnetism, and uniaxial AMF by more than 30%
A survey on active simultaneous localization and mapping: state of the art and new frontiers
Active simultaneous localization and mapping (SLAM) is the problem of planning and controlling the motion of a robot to build the most accurate and complete model of the surrounding environment. Since the first foundational work in active perception appeared, more than three decades ago, this field has received increasing attention across different scientific communities. This has brought about many different approaches and formulations, and makes a review of the current trends necessary and extremely valuable for both new and experienced researchers. In this article, we survey the state of the art in active SLAM and take an in-depth look at the open challenges that still require attention to meet the needs of modern applications. After providing a historical perspective, we present a unified problem formulation and review the well-established modular solution scheme, which decouples the problem into three stages that identify, select, and execute potential navigation actions. We then analyze alternative approaches, including belief-space planning and deep reinforcement learning techniques, and review related work on multirobot coordination. This article concludes with a discussion of new research directions, addressing reproducible research, active spatial perception, and practical applications, among other topics
Robust state estimation methods for robotics applications
State estimation is an integral component of any autonomous robotic system. Finding the correct position, velocity, and orientation of an agent in its environment enables it to do other tasks like mapping and interacting with the environment, and collaborating with other agents. State estimation is achieved by using data obtained from multiple sensors and fusing them in a probabilistic framework. These include inertial data from Inertial Measurement Unit (IMU), images from camera, range data from lidars, and positioning data from Global Navigation Satellite Systems (GNSS) receivers. The main challenge faced in sensor-based state estimation is the presence of noisy, erroneous, and even lack of informative data. Some common examples of such situations include wrong feature matching between images or point clouds, false loop-closures due to perceptual aliasing (different places that look similar can confuse the robot), presence of dynamic objects in the environment (odometry algorithms assume a static environment), multipath errors for GNSS (signals for satellites jumping off tall structures like buildings before reaching receivers) and more. This work studies existing and new ways of how standard estimation algorithms like the Kalman filter and factor graphs can be made robust to such adverse conditions without losing performance in ideal outlier-free conditions. The first part of this work demonstrates the importance of robust Kalman filters on wheel-inertial odometry for high-slip terrain. Next, inertial data is integrated into GNSS factor graphs to improve the accuracy and robustness of GNSS factor graphs. Lastly, a combined framework for improving the robustness of non-linear least squares and estimating the inlier noise threshold is proposed and tested with point cloud registration and lidar-inertial odometry algorithms followed by an algorithmic analysis of optimizing generalized robust cost functions with factor graphs for GNSS positioning problem
Walking Your LiDOG: A Journey Through Multiple Domains for LiDAR Semantic Segmentation
The ability to deploy robots that can operate safely in diverse environments
is crucial for developing embodied intelligent agents. As a community, we have
made tremendous progress in within-domain LiDAR semantic segmentation. However,
do these methods generalize across domains? To answer this question, we design
the first experimental setup for studying domain generalization (DG) for LiDAR
semantic segmentation (DG-LSS). Our results confirm a significant gap between
methods, evaluated in a cross-domain setting: for example, a model trained on
the source dataset (SemanticKITTI) obtains mIoU on the target data,
compared to mIoU obtained by the model trained on the target domain
(nuScenes). To tackle this gap, we propose the first method specifically
designed for DG-LSS, which obtains mIoU on the target domain,
outperforming all baselines. Our method augments a sparse-convolutional
encoder-decoder 3D segmentation network with an additional, dense 2D
convolutional decoder that learns to classify a birds-eye view of the point
cloud. This simple auxiliary task encourages the 3D network to learn features
that are robust to sensor placement shifts and resolution, and are transferable
across domains. With this work, we aim to inspire the community to develop and
evaluate future models in such cross-domain conditions.Comment: Accepted at ICCV 202
Sensor Fusion and Deep Learning for Indoor Agent Localization
Autonomous, self-navigating agents have been rising in popularity due to a push for a more technologically aided future. From cars to vacuum cleaners, the applications of self-navigating agents are vast and span many different fields and aspects of life. As the demand for these autonomous robotic agents has been increasing, so has the demand for innovative features, robust behavior, and lower cost hardware. One particular area with a constant demand for improvement is localization, or an agent\u27s ability to determine where it is located within its environment. Whether the agent\u27s environment is primarily indoor or outdoor, dense or sparse, static or dynamic, an agent must be able to have knowledge of its location. Many different techniques exist today for localization, each having its strengths and weaknesses. Despite the abundance of different techniques, there is still room for improvement. This research presents a novel indoor localization algorithm that fuses data from multiple sensors for a relatively low cost. Inspired by recent innovations in deep learning and particle filters, a fast, robust, and accurate autonomous localization system has been created. Results demonstrate that the proposed system is both real-time and robust against changing conditions within the environment
Recommended from our members
Multi-SLAM Systems for Fault-Tolerant Simultaneous Localization and Mapping
Mobile robots need accurate, high fidelity models of their operating environments in order to complete their tasks safely and efficiently. Generating these models is most often done via Simultaneous Localization and Mapping (SLAM), a paradigm where the robot alternatively estimates the most up-to-date model of the environment and its position relative to this model as it acquires new information from its sensors over time. Because robots operate in many different environments with different compute, memory, sensing, and form constraints, the nature and quality of information available to individual instances of different SLAM systems varies substantially. `One-size-fits-all\u27 solutions are thus exceedingly difficult to engineer, and highly specialized systems, which represent the state-of-the-art for most types of deployments, are not robust to operating conditions in which their assumptions are not met. This thesis seeks to investigate an alternative approach to these robustness and universality problems by incorporating existing SLAM solutions within a larger framework supported by planning and learning. The central idea is to combine learned models that estimate SLAM algorithm performance under a variety of sensory conditions, in this case neural networks, with planners designed for planning under uncertainty and partial observability, in this case partially observable Markov decision problems (POMDPs). Models of existing SLAM algorithms can be learned, and these models can then be used online to estimate the performance of a range of solutions to the SLAM problem at hand. The POMDP policy then selects the appropriate algorithm, given the estimated performance, cost of switching methods, and other information. This general approach may also be applicable to many other robotics problems that rely on data-fusion, such as grasp planning, motion planning, or object identification
A Survey on Aerial Swarm Robotics
The use of aerial swarms to solve real-world problems has been increasing steadily, accompanied by falling prices and improving performance of communication, sensing, and processing hardware. The commoditization of hardware has reduced unit costs, thereby lowering the barriers to entry to the field of aerial swarm robotics. A key enabling technology for swarms is the family of algorithms that allow the individual members of the swarm to communicate and allocate tasks amongst themselves, plan their trajectories, and coordinate their flight in such a way that the overall objectives of the swarm are achieved efficiently. These algorithms, often organized in a hierarchical fashion, endow the swarm with autonomy at every level, and the role of a human operator can be reduced, in principle, to interactions at a higher level without direct intervention. This technology depends on the clever and innovative application of theoretical tools from control and estimation. This paper reviews the state of the art of these theoretical tools, specifically focusing on how they have been developed for, and applied to, aerial swarms. Aerial swarms differ from swarms of ground-based vehicles in two respects: they operate in a three-dimensional space and the dynamics of individual vehicles adds an extra layer of complexity. We review dynamic modeling and conditions for stability and controllability that are essential in order to achieve cooperative flight and distributed sensing. The main sections of this paper focus on major results covering trajectory generation, task allocation, adversarial control, distributed sensing, monitoring, and mapping. Wherever possible, we indicate how the physics and subsystem technologies of aerial robots are brought to bear on these individual areas
- …