57,526 research outputs found
Safe Local Exploration for Replanning in Cluttered Unknown Environments for Micro-Aerial Vehicles
In order to enable Micro-Aerial Vehicles (MAVs) to assist in complex,
unknown, unstructured environments, they must be able to navigate with
guaranteed safety, even when faced with a cluttered environment they have no
prior knowledge of. While trajectory optimization-based local planners have
been shown to perform well in these cases, prior work either does not address
how to deal with local minima in the optimization problem, or solves it by
using an optimistic global planner.
We present a conservative trajectory optimization-based local planner,
coupled with a local exploration strategy that selects intermediate goals. We
perform extensive simulations to show that this system performs better than the
standard approach of using an optimistic global planner, and also outperforms
doing a single exploration step when the local planner is stuck. The method is
validated through experiments in a variety of highly cluttered environments
including a dense forest. These experiments show the complete system running in
real time fully onboard an MAV, mapping and replanning at 4 Hz.Comment: Accepted to ICRA 2018 and RA-L 201
Multi-criteria Evolution of Neural Network Topologies: Balancing Experience and Performance in Autonomous Systems
Majority of Artificial Neural Network (ANN) implementations in autonomous
systems use a fixed/user-prescribed network topology, leading to sub-optimal
performance and low portability. The existing neuro-evolution of augmenting
topology or NEAT paradigm offers a powerful alternative by allowing the network
topology and the connection weights to be simultaneously optimized through an
evolutionary process. However, most NEAT implementations allow the
consideration of only a single objective. There also persists the question of
how to tractably introduce topological diversification that mitigates
overfitting to training scenarios. To address these gaps, this paper develops a
multi-objective neuro-evolution algorithm. While adopting the basic elements of
NEAT, important modifications are made to the selection, speciation, and
mutation processes. With the backdrop of small-robot path-planning
applications, an experience-gain criterion is derived to encapsulate the amount
of diverse local environment encountered by the system. This criterion
facilitates the evolution of genes that support exploration, thereby seeking to
generalize from a smaller set of mission scenarios than possible with
performance maximization alone. The effectiveness of the single-objective
(optimizing performance) and the multi-objective (optimizing performance and
experience-gain) neuro-evolution approaches are evaluated on two different
small-robot cases, with ANNs obtained by the multi-objective optimization
observed to provide superior performance in unseen scenarios
Automating Vehicles by Deep Reinforcement Learning using Task Separation with Hill Climbing
Within the context of autonomous driving a model-based reinforcement learning
algorithm is proposed for the design of neural network-parameterized
controllers. Classical model-based control methods, which include sampling- and
lattice-based algorithms and model predictive control, suffer from the
trade-off between model complexity and computational burden required for the
online solution of expensive optimization or search problems at every short
sampling time. To circumvent this trade-off, a 2-step procedure is motivated:
first learning of a controller during offline training based on an arbitrarily
complicated mathematical system model, before online fast feedforward
evaluation of the trained controller. The contribution of this paper is the
proposition of a simple gradient-free and model-based algorithm for deep
reinforcement learning using task separation with hill climbing (TSHC). In
particular, (i) simultaneous training on separate deterministic tasks with the
purpose of encoding many motion primitives in a neural network, and (ii) the
employment of maximally sparse rewards in combination with virtual velocity
constraints (VVCs) in setpoint proximity are advocated.Comment: 10 pages, 6 figures, 1 tabl
Numerical Efficiency of Inverse Simulation Methods Applied to a Wheeled Rover
Extending the navigational capability of planetary rovers is essential for increasing the scientific outputs from such exploratory missions. In this paper a navigation method based on Inverse Simulation is applied to a four wheel rover. The method calculates the required control inputs to achieve a desired, specified response. Here this is a desired trajectory defined as a series of waypoints. Inverse Simulation considers the complete system dynamics of the rover to calculate the control input using an iterative, numerical Newton - Raphson scheme. The paper provides an insight into the numerical parameters that affect the performance of the method. Also, the influence of varying the timestep and the convergence tolerance is examined in terms of the quality of the calculated control input and the resulting trajectory, as well as the execution time. From this analysis a set of parameters and recommendations to successfully apply Inverse Simulation to a rover is presented
A biologically inspired meta-control navigation system for the Psikharpax rat robot
A biologically inspired navigation system for the mobile rat-like robot named Psikharpax is presented, allowing for self-localization and autonomous navigation in an initially unknown environment. The ability of parts of the model (e. g. the strategy selection mechanism) to reproduce rat behavioral data in various maze tasks has been validated before in simulations. But the capacity of the model to work on a real robot platform had not been tested. This paper presents our work on the implementation on the Psikharpax robot of two independent navigation strategies (a place-based planning strategy and a cue-guided taxon strategy) and a strategy selection meta-controller. We show how our robot can memorize which was the optimal strategy in each situation, by means of a reinforcement learning algorithm. Moreover, a context detector enables the controller to quickly adapt to changes in the environment-recognized as new contexts-and to restore previously acquired strategy preferences when a previously experienced context is recognized. This produces adaptivity closer to rat behavioral performance and constitutes a computational proposition of the role of the rat prefrontal cortex in strategy shifting. Moreover, such a brain-inspired meta-controller may provide an advancement for learning architectures in robotics
- …