2,867 research outputs found
Topomap: Topological Mapping and Navigation Based on Visual SLAM Maps
Visual robot navigation within large-scale, semi-structured environments
deals with various challenges such as computation intensive path planning
algorithms or insufficient knowledge about traversable spaces. Moreover, many
state-of-the-art navigation approaches only operate locally instead of gaining
a more conceptual understanding of the planning objective. This limits the
complexity of tasks a robot can accomplish and makes it harder to deal with
uncertainties that are present in the context of real-time robotics
applications. In this work, we present Topomap, a framework which simplifies
the navigation task by providing a map to the robot which is tailored for path
planning use. This novel approach transforms a sparse feature-based map from a
visual Simultaneous Localization And Mapping (SLAM) system into a
three-dimensional topological map. This is done in two steps. First, we extract
occupancy information directly from the noisy sparse point cloud. Then, we
create a set of convex free-space clusters, which are the vertices of the
topological map. We show that this representation improves the efficiency of
global planning, and we provide a complete derivation of our algorithm.
Planning experiments on real world datasets demonstrate that we achieve similar
performance as RRT* with significantly lower computation times and storage
requirements. Finally, we test our algorithm on a mobile robotic platform to
prove its advantages.Comment: 8 page
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
Fast, Accurate Thin-Structure Obstacle Detection for Autonomous Mobile Robots
Safety is paramount for mobile robotic platforms such as self-driving cars
and unmanned aerial vehicles. This work is devoted to a task that is
indispensable for safety yet was largely overlooked in the past -- detecting
obstacles that are of very thin structures, such as wires, cables and tree
branches. This is a challenging problem, as thin objects can be problematic for
active sensors such as lidar and sonar and even for stereo cameras. In this
work, we propose to use video sequences for thin obstacle detection. We
represent obstacles with edges in the video frames, and reconstruct them in 3D
using efficient edge-based visual odometry techniques. We provide both a
monocular camera solution and a stereo camera solution. The former incorporates
Inertial Measurement Unit (IMU) data to solve scale ambiguity, while the latter
enjoys a novel, purely vision-based solution. Experiments demonstrated that the
proposed methods are fast and able to detect thin obstacles robustly and
accurately under various conditions.Comment: Appeared at IEEE CVPR 2017 Workshop on Embedded Visio
C-blox: A Scalable and Consistent TSDF-based Dense Mapping Approach
In many applications, maintaining a consistent dense map of the environment
is key to enabling robotic platforms to perform higher level decision making.
Several works have addressed the challenge of creating precise dense 3D maps
from visual sensors providing depth information. However, during operation over
longer missions, reconstructions can easily become inconsistent due to
accumulated camera tracking error and delayed loop closure. Without explicitly
addressing the problem of map consistency, recovery from such distortions tends
to be difficult. We present a novel system for dense 3D mapping which addresses
the challenge of building consistent maps while dealing with scalability.
Central to our approach is the representation of the environment as a
collection of overlapping TSDF subvolumes. These subvolumes are localized
through feature-based camera tracking and bundle adjustment. Our main
contribution is a pipeline for identifying stable regions in the map, and to
fuse the contributing subvolumes. This approach allows us to reduce map growth
while still maintaining consistency. We demonstrate the proposed system on a
publicly available dataset and simulation engine, and demonstrate the efficacy
of the proposed approach for building consistent and scalable maps. Finally we
demonstrate our approach running in real-time on-board a lightweight MAV.Comment: 8 pages, 5 figures, conferenc
LDSO: Direct Sparse Odometry with Loop Closure
In this paper we present an extension of Direct Sparse Odometry (DSO) to a
monocular visual SLAM system with loop closure detection and pose-graph
optimization (LDSO). As a direct technique, DSO can utilize any image pixel
with sufficient intensity gradient, which makes it robust even in featureless
areas. LDSO retains this robustness, while at the same time ensuring
repeatability of some of these points by favoring corner features in the
tracking frontend. This repeatability allows to reliably detect loop closure
candidates with a conventional feature-based bag-of-words (BoW) approach. Loop
closure candidates are verified geometrically and Sim(3) relative pose
constraints are estimated by jointly minimizing 2D and 3D geometric error
terms. These constraints are fused with a co-visibility graph of relative poses
extracted from DSO's sliding window optimization. Our evaluation on publicly
available datasets demonstrates that the modified point selection strategy
retains the tracking accuracy and robustness, and the integrated pose-graph
optimization significantly reduces the accumulated rotation-, translation- and
scale-drift, resulting in an overall performance comparable to state-of-the-art
feature-based systems, even without global bundle adjustment
- …