5,039 research outputs found
Topomap: Topological Mapping and Navigation Based on Visual SLAM Maps
Visual robot navigation within large-scale, semi-structured environments
deals with various challenges such as computation intensive path planning
algorithms or insufficient knowledge about traversable spaces. Moreover, many
state-of-the-art navigation approaches only operate locally instead of gaining
a more conceptual understanding of the planning objective. This limits the
complexity of tasks a robot can accomplish and makes it harder to deal with
uncertainties that are present in the context of real-time robotics
applications. In this work, we present Topomap, a framework which simplifies
the navigation task by providing a map to the robot which is tailored for path
planning use. This novel approach transforms a sparse feature-based map from a
visual Simultaneous Localization And Mapping (SLAM) system into a
three-dimensional topological map. This is done in two steps. First, we extract
occupancy information directly from the noisy sparse point cloud. Then, we
create a set of convex free-space clusters, which are the vertices of the
topological map. We show that this representation improves the efficiency of
global planning, and we provide a complete derivation of our algorithm.
Planning experiments on real world datasets demonstrate that we achieve similar
performance as RRT* with significantly lower computation times and storage
requirements. Finally, we test our algorithm on a mobile robotic platform to
prove its advantages.Comment: 8 page
Recommended from our members
A corpus-based analysis of route instructions in human-robot interaction
This paper investigates how users employ spatial descriptions to navigate a speech-enabled robot. We created a simulated environment in which users gave route instructions in a dialogic real-time interaction with a robot, which was
operated by naĂŻve participants. The ability of robot monitoring was also manipulated in two experimental conditions. The results provide evidence that the content of the instructions and strategies of the users vary depending on the conditions and
demands of the interaction. As expected, the route instructions frequently were underspecified and arbitrary. The findings of
this study elucidate the complexity in interpreting spatial language in HRI. However, they also point to the need for
endowing mobile robots with richer dialogue resources to compensate for the uncertainties arising from language as well
as the environment
Characterization of image sets: the Galois Lattice approach
This paper presents a new method for supervised image
classification. One or several landmarks are attached to each class, with the intention of characterizing it and discriminating it from the other classes. The different features, deduced from image primitives, and their relationships with the sets of images are structured and organized into a hierarchy thanks to an original method relying on a mathematical formalism called Galois (or Concept) Lattices. Such lattices allow us to select features as landmarks of specific classes. This paper details the feature selection process and illustrates this through a robotic example in a structured environment. The class of any image is the room from which the image is shot by the robot camera. In the discussion, we compare this approach with decision trees and we give some issues for future research
Technical report on Optimization-Based Bearing-Only Visual Homing with Applications to a 2-D Unicycle Model
We consider the problem of bearing-based visual homing: Given a mobile robot
which can measure bearing directions with respect to known landmarks, the goal
is to guide the robot toward a desired "home" location. We propose a control
law based on the gradient field of a Lyapunov function, and give sufficient
conditions for global convergence. We show that the well-known Average Landmark
Vector method (for which no convergence proof was known) can be obtained as a
particular case of our framework. We then derive a sliding mode control law for
a unicycle model which follows this gradient field. Both controllers do not
depend on range information. Finally, we also show how our framework can be
used to characterize the sensitivity of a home location with respect to noise
in the specified bearings. This is an extended version of the conference paper
[1].Comment: This is an extender version of R. Tron and K. Daniilidis, "An
optimization approach to bearing-only visual homing with applications to a
2-D unicycle model," in IEEE International Conference on Robotics and
Automation, 2014, containing additional proof
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
Simultaneous localization and map-building using active vision
An active approach to sensing can provide the focused measurement capability over a wide field of view which allows correctly formulated Simultaneous Localization and Map-Building (SLAM) to be implemented with vision, permitting repeatable long-term localization using only naturally occurring, automatically-detected features. In this paper, we present the first example of a general system for autonomous localization using active vision, enabled here by a high-performance stereo head, addressing such issues as uncertainty-based measurement selection, automatic map-maintenance, and goal-directed steering. We present varied real-time experiments in a complex environment.Published versio
Robust Legged Robot State Estimation Using Factor Graph Optimization
Legged robots, specifically quadrupeds, are becoming increasingly attractive
for industrial applications such as inspection. However, to leave the
laboratory and to become useful to an end user requires reliability in harsh
conditions. From the perspective of state estimation, it is essential to be
able to accurately estimate the robot's state despite challenges such as uneven
or slippery terrain, textureless and reflective scenes, as well as dynamic
camera occlusions. We are motivated to reduce the dependency on foot contact
classifications, which fail when slipping, and to reduce position drift during
dynamic motions such as trotting. To this end, we present a factor graph
optimization method for state estimation which tightly fuses and smooths
inertial navigation, leg odometry and visual odometry. The effectiveness of the
approach is demonstrated using the ANYmal quadruped robot navigating in a
realistic outdoor industrial environment. This experiment included trotting,
walking, crossing obstacles and ascending a staircase. The proposed approach
decreased the relative position error by up to 55% and absolute position error
by 76% compared to kinematic-inertial odometry.Comment: 8 pages, 12 figures. Accepted to RA-L + IROS 2019, July 201
Attention and Anticipation in Fast Visual-Inertial Navigation
We study a Visual-Inertial Navigation (VIN) problem in which a robot needs to
estimate its state using an on-board camera and an inertial sensor, without any
prior knowledge of the external environment. We consider the case in which the
robot can allocate limited resources to VIN, due to tight computational
constraints. Therefore, we answer the following question: under limited
resources, what are the most relevant visual cues to maximize the performance
of visual-inertial navigation? Our approach has four key ingredients. First, it
is task-driven, in that the selection of the visual cues is guided by a metric
quantifying the VIN performance. Second, it exploits the notion of
anticipation, since it uses a simplified model for forward-simulation of robot
dynamics, predicting the utility of a set of visual cues over a future time
horizon. Third, it is efficient and easy to implement, since it leads to a
greedy algorithm for the selection of the most relevant visual cues. Fourth, it
provides formal performance guarantees: we leverage submodularity to prove that
the greedy selection cannot be far from the optimal (combinatorial) selection.
Simulations and real experiments on agile drones show that our approach ensures
state-of-the-art VIN performance while maintaining a lean processing time. In
the easy scenarios, our approach outperforms appearance-based feature selection
in terms of localization errors. In the most challenging scenarios, it enables
accurate visual-inertial navigation while appearance-based feature selection
fails to track robot's motion during aggressive maneuvers.Comment: 20 pages, 7 figures, 2 table
- âŠ