335 research outputs found
Non-Linear Model Predictive Control with Adaptive Time-Mesh Refinement
In this paper, we present a novel solution for real-time, Non-Linear Model
Predictive Control (NMPC) exploiting a time-mesh refinement strategy. The
proposed controller formulates the Optimal Control Problem (OCP) in terms of
flat outputs over an adaptive lattice. In common approximated OCP solutions,
the number of discretization points composing the lattice represents a critical
upper bound for real-time applications. The proposed NMPC-based technique
refines the initially uniform time horizon by adding time steps with a sampling
criterion that aims to reduce the discretization error. This enables a higher
accuracy in the initial part of the receding horizon, which is more relevant to
NMPC, while keeping bounded the number of discretization points. By combining
this feature with an efficient Least Square formulation, our solver is also
extremely time-efficient, generating trajectories of multiple seconds within
only a few milliseconds. The performance of the proposed approach has been
validated in a high fidelity simulation environment, by using an UAV platform.
We also released our implementation as open source C++ code.Comment: In: 2018 IEEE International Conference on Simulation, Modeling, and
Programming for Autonomous Robots (SIMPAR 2018
A Survey of path following control strategies for UAVs focused on quadrotors
The trajectory control problem, defined as making a vehicle follow a pre-established path in space, can be solved by means of trajectory tracking or path following. In the trajectory tracking problem a timed reference position is tracked. The path following approach removes any time dependence of the problem, resulting in many advantages on the control performance and design. An exhaustive review of path following algorithms applied to quadrotor vehicles has been carried out, the most relevant are studied in this paper. Then, four of these algorithms have been implemented and compared in a quadrotor simulation platform: Backstepping and Feedback Linearisation control-oriented algorithms and NLGL and Carrot-Chasing geometric algorithms.Peer ReviewedPostprint (author's final draft
Suspended Load Path Tracking Control Using a Tilt-rotor UAV Based on Zonotopic State Estimation
This work addresses the problem of path tracking control of a suspended load
using a tilt-rotor UAV. The main challenge in controlling this kind of system
arises from the dynamic behavior imposed by the load, which is usually coupled
to the UAV by means of a rope, adding unactuated degrees of freedom to the
whole system. Furthermore, to perform the load transportation it is often
needed the knowledge of the load position to accomplish the task. Since
available sensors are commonly embedded in the mobile platform, information on
the load position may not be directly available. To solve this problem in this
work, initially, the kinematics of the multi-body mechanical system are
formulated from the load's perspective, from which a detailed dynamic model is
derived using the Euler-Lagrange approach, yielding a highly coupled, nonlinear
state-space representation of the system, affine in the inputs, with the load's
position and orientation directly represented by state variables. A zonotopic
state estimator is proposed to solve the problem of estimating the load
position and orientation, which is formulated based on sensors located at the
aircraft, with different sampling times, and unknown-but-bounded measurement
noise. To solve the path tracking problem, a discrete-time mixed
controller with pole-placement constraints
is designed with guaranteed time-response properties and robust to unmodeled
dynamics, parametric uncertainties, and external disturbances. Results from
numerical experiments, performed in a platform based on the Gazebo simulator
and on a Computer Aided Design (CAD) model of the system, are presented to
corroborate the performance of the zonotopic state estimator along with the
designed controller
Human-in-the-Loop Methods for Data-Driven and Reinforcement Learning Systems
Recent successes combine reinforcement learning algorithms and deep neural
networks, despite reinforcement learning not being widely applied to robotics
and real world scenarios. This can be attributed to the fact that current
state-of-the-art, end-to-end reinforcement learning approaches still require
thousands or millions of data samples to converge to a satisfactory policy and
are subject to catastrophic failures during training. Conversely, in real world
scenarios and after just a few data samples, humans are able to either provide
demonstrations of the task, intervene to prevent catastrophic actions, or
simply evaluate if the policy is performing correctly. This research
investigates how to integrate these human interaction modalities to the
reinforcement learning loop, increasing sample efficiency and enabling
real-time reinforcement learning in robotics and real world scenarios. This
novel theoretical foundation is called Cycle-of-Learning, a reference to how
different human interaction modalities, namely, task demonstration,
intervention, and evaluation, are cycled and combined to reinforcement learning
algorithms. Results presented in this work show that the reward signal that is
learned based upon human interaction accelerates the rate of learning of
reinforcement learning algorithms and that learning from a combination of human
demonstrations and interventions is faster and more sample efficient when
compared to traditional supervised learning algorithms. Finally,
Cycle-of-Learning develops an effective transition between policies learned
using human demonstrations and interventions to reinforcement learning. The
theoretical foundation developed by this research opens new research paths to
human-agent teaming scenarios where autonomous agents are able to learn from
human teammates and adapt to mission performance metrics in real-time and in
real world scenarios.Comment: PhD thesis, Aerospace Engineering, Texas A&M (2020). For more
information, see https://vggoecks.com
An Uncertainty Quantification Framework for Autonomous System Tracking and Health Monitoring
This work proposes a perspective towards establishing a framework for uncertainty quantification of autonomous system tracking and health monitoring. The approach leverages the use of a predictive process structure, which maps uncertainty sources and their interaction according to the quantity of interest and the goal of the predictive estimation. It is systematic and uses basic elements that are system agnostic, and therefore needs to be tailored according to the specificity of the application. This work is motivated by the interest in low-altitude unmanned aerial vehicle operations, where awareness of vehicle and airspace state becomes more relevant as the density of autonomous operations grows rapidly. Predicted scenarios in the area of small vehicle operations and urban air mobility have no precedent, and holistic frameworks to perform prognostics and health management (PHM) at the system- and airspace-level are missing formal approaches to account for uncertainty. At the end of the paper, two case studies demonstrate implementation framework of trajectory tracking and health diagnosis for a small unmanned aerial vehicle
Neuromorphic Systems for Pattern Recognition and Uav Trajectory Planning
Detection and control are two essential components in an intelligent system. This thesis investigates novel techniques in both areas with a focus on the applications of handwritten text recognition and UAV flight control. Recognizing handwritten texts is a challenging task due to many different writing styles and lack of clear boundary between adjacent characters. The difficulty is greatly increased if the detection algorithms is solely based on pattern matching without information of dynamics of handwriting trajectories. Motivated by the aforementioned challenges, this thesis first investigates the pattern recognition problem. We use offline handwritten texts recognition as a case study to explore the performance of a recurrent belief propagation model. We first develop a probabilistic inference network to post process the recognition results of deep Convolutional Neural Network (CNN) (e.g. LeNet) and collect individual characters to form words. The output of the inference network is a set of words and their probability. A series of post processing and improvement techniques are then introduced to further increase the recognition accuracy. We study the performance of proposed model through various comparisons. The results show that it significantly improves the accuracy by correcting deletion, insertion and replacement errors, which are the main sources of invalid candidate words.
Deep Reinforcement Learning (DRL) has widely been applied to control the autonomous systems because it provides solutions for various complex decision-making tasks that previously could not be solved solely with deep learning. To enable autonomous Unmanned Aerial Vehicles (UAV), this thesis presents a two-level trajectory planning framework for UAVs in an indoor environment. A sequence of waypoints is selected at the higher-level, which leads the UAV from its current position to the destination. At the lower-level, an optimal trajectory is generated analytically between each pair of adjacent waypoints. The goal of trajectory generation is to maintain the stability of the UAV, and the goal of the waypoints planning is to select waypoints with the lowest control thrust throughout the entire trip while avoiding collisions with obstacles. The entire framework is implemented using DRL, which learns the highly complicated and nonlinear interaction between those two levels, and the impact from the environment. Given the pre-planned trajectory, this thesis further presents an actor-critic reinforcement learning framework that realizes continuous trajectory control of the UAV through a set of desired waypoints. We construct a deep neural network and develop reinforcement learning for better trajectory tracking. In addition, Field Programmable Gate Arrays (FPGA) based hardware acceleration is designed for energy efficient real-time control.
If we are to integrate the trajectory planning model onto a UAV system for real-time on-board planning, a key challenge is how to deliver required performance under strict memory and computational constraints. Techniques that compress Deep Neural Network (DNN) models attract our attention because they allow optimized neural network models to be efficiently deployed on platforms with limited energy and storage capacity. However, conventional model compression techniques prune the DNN after it is fully trained, which is very time-consuming especially when the model is trained using DRL. To overcome the limitation, we present an early phase integrated neural network weight compression system for DRL based waypoints planning. By applying pruning at an early phase, the compression of the DRL model can be realized without significant overhead in training. By tightly integrating pruning and retraining at the early phase, we achieve a higher model compression rate, reduce more memory and computing complexity, and improve the success rate compared to the original work
- …