335 research outputs found

    Non-Linear Model Predictive Control with Adaptive Time-Mesh Refinement

    Full text link
    In this paper, we present a novel solution for real-time, Non-Linear Model Predictive Control (NMPC) exploiting a time-mesh refinement strategy. The proposed controller formulates the Optimal Control Problem (OCP) in terms of flat outputs over an adaptive lattice. In common approximated OCP solutions, the number of discretization points composing the lattice represents a critical upper bound for real-time applications. The proposed NMPC-based technique refines the initially uniform time horizon by adding time steps with a sampling criterion that aims to reduce the discretization error. This enables a higher accuracy in the initial part of the receding horizon, which is more relevant to NMPC, while keeping bounded the number of discretization points. By combining this feature with an efficient Least Square formulation, our solver is also extremely time-efficient, generating trajectories of multiple seconds within only a few milliseconds. The performance of the proposed approach has been validated in a high fidelity simulation environment, by using an UAV platform. We also released our implementation as open source C++ code.Comment: In: 2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR 2018

    A Survey of path following control strategies for UAVs focused on quadrotors

    Get PDF
    The trajectory control problem, defined as making a vehicle follow a pre-established path in space, can be solved by means of trajectory tracking or path following. In the trajectory tracking problem a timed reference position is tracked. The path following approach removes any time dependence of the problem, resulting in many advantages on the control performance and design. An exhaustive review of path following algorithms applied to quadrotor vehicles has been carried out, the most relevant are studied in this paper. Then, four of these algorithms have been implemented and compared in a quadrotor simulation platform: Backstepping and Feedback Linearisation control-oriented algorithms and NLGL and Carrot-Chasing geometric algorithms.Peer ReviewedPostprint (author's final draft

    Suspended Load Path Tracking Control Using a Tilt-rotor UAV Based on Zonotopic State Estimation

    Full text link
    This work addresses the problem of path tracking control of a suspended load using a tilt-rotor UAV. The main challenge in controlling this kind of system arises from the dynamic behavior imposed by the load, which is usually coupled to the UAV by means of a rope, adding unactuated degrees of freedom to the whole system. Furthermore, to perform the load transportation it is often needed the knowledge of the load position to accomplish the task. Since available sensors are commonly embedded in the mobile platform, information on the load position may not be directly available. To solve this problem in this work, initially, the kinematics of the multi-body mechanical system are formulated from the load's perspective, from which a detailed dynamic model is derived using the Euler-Lagrange approach, yielding a highly coupled, nonlinear state-space representation of the system, affine in the inputs, with the load's position and orientation directly represented by state variables. A zonotopic state estimator is proposed to solve the problem of estimating the load position and orientation, which is formulated based on sensors located at the aircraft, with different sampling times, and unknown-but-bounded measurement noise. To solve the path tracking problem, a discrete-time mixed H2/H\mathcal{H}_2/\mathcal{H}_\infty controller with pole-placement constraints is designed with guaranteed time-response properties and robust to unmodeled dynamics, parametric uncertainties, and external disturbances. Results from numerical experiments, performed in a platform based on the Gazebo simulator and on a Computer Aided Design (CAD) model of the system, are presented to corroborate the performance of the zonotopic state estimator along with the designed controller

    Human-in-the-Loop Methods for Data-Driven and Reinforcement Learning Systems

    Get PDF
    Recent successes combine reinforcement learning algorithms and deep neural networks, despite reinforcement learning not being widely applied to robotics and real world scenarios. This can be attributed to the fact that current state-of-the-art, end-to-end reinforcement learning approaches still require thousands or millions of data samples to converge to a satisfactory policy and are subject to catastrophic failures during training. Conversely, in real world scenarios and after just a few data samples, humans are able to either provide demonstrations of the task, intervene to prevent catastrophic actions, or simply evaluate if the policy is performing correctly. This research investigates how to integrate these human interaction modalities to the reinforcement learning loop, increasing sample efficiency and enabling real-time reinforcement learning in robotics and real world scenarios. This novel theoretical foundation is called Cycle-of-Learning, a reference to how different human interaction modalities, namely, task demonstration, intervention, and evaluation, are cycled and combined to reinforcement learning algorithms. Results presented in this work show that the reward signal that is learned based upon human interaction accelerates the rate of learning of reinforcement learning algorithms and that learning from a combination of human demonstrations and interventions is faster and more sample efficient when compared to traditional supervised learning algorithms. Finally, Cycle-of-Learning develops an effective transition between policies learned using human demonstrations and interventions to reinforcement learning. The theoretical foundation developed by this research opens new research paths to human-agent teaming scenarios where autonomous agents are able to learn from human teammates and adapt to mission performance metrics in real-time and in real world scenarios.Comment: PhD thesis, Aerospace Engineering, Texas A&M (2020). For more information, see https://vggoecks.com

    An Uncertainty Quantification Framework for Autonomous System Tracking and Health Monitoring

    Get PDF
    This work proposes a perspective towards establishing a framework for uncertainty quantification of autonomous system tracking and health monitoring. The approach leverages the use of a predictive process structure, which maps uncertainty sources and their interaction according to the quantity of interest and the goal of the predictive estimation. It is systematic and uses basic elements that are system agnostic, and therefore needs to be tailored according to the specificity of the application. This work is motivated by the interest in low-altitude unmanned aerial vehicle operations, where awareness of vehicle and airspace state becomes more relevant as the density of autonomous operations grows rapidly. Predicted scenarios in the area of small vehicle operations and urban air mobility have no precedent, and holistic frameworks to perform prognostics and health management (PHM) at the system- and airspace-level are missing formal approaches to account for uncertainty. At the end of the paper, two case studies demonstrate implementation framework of trajectory tracking and health diagnosis for a small unmanned aerial vehicle

    Neuromorphic Systems for Pattern Recognition and Uav Trajectory Planning

    Get PDF
    Detection and control are two essential components in an intelligent system. This thesis investigates novel techniques in both areas with a focus on the applications of handwritten text recognition and UAV flight control. Recognizing handwritten texts is a challenging task due to many different writing styles and lack of clear boundary between adjacent characters. The difficulty is greatly increased if the detection algorithms is solely based on pattern matching without information of dynamics of handwriting trajectories. Motivated by the aforementioned challenges, this thesis first investigates the pattern recognition problem. We use offline handwritten texts recognition as a case study to explore the performance of a recurrent belief propagation model. We first develop a probabilistic inference network to post process the recognition results of deep Convolutional Neural Network (CNN) (e.g. LeNet) and collect individual characters to form words. The output of the inference network is a set of words and their probability. A series of post processing and improvement techniques are then introduced to further increase the recognition accuracy. We study the performance of proposed model through various comparisons. The results show that it significantly improves the accuracy by correcting deletion, insertion and replacement errors, which are the main sources of invalid candidate words. Deep Reinforcement Learning (DRL) has widely been applied to control the autonomous systems because it provides solutions for various complex decision-making tasks that previously could not be solved solely with deep learning. To enable autonomous Unmanned Aerial Vehicles (UAV), this thesis presents a two-level trajectory planning framework for UAVs in an indoor environment. A sequence of waypoints is selected at the higher-level, which leads the UAV from its current position to the destination. At the lower-level, an optimal trajectory is generated analytically between each pair of adjacent waypoints. The goal of trajectory generation is to maintain the stability of the UAV, and the goal of the waypoints planning is to select waypoints with the lowest control thrust throughout the entire trip while avoiding collisions with obstacles. The entire framework is implemented using DRL, which learns the highly complicated and nonlinear interaction between those two levels, and the impact from the environment. Given the pre-planned trajectory, this thesis further presents an actor-critic reinforcement learning framework that realizes continuous trajectory control of the UAV through a set of desired waypoints. We construct a deep neural network and develop reinforcement learning for better trajectory tracking. In addition, Field Programmable Gate Arrays (FPGA) based hardware acceleration is designed for energy efficient real-time control. If we are to integrate the trajectory planning model onto a UAV system for real-time on-board planning, a key challenge is how to deliver required performance under strict memory and computational constraints. Techniques that compress Deep Neural Network (DNN) models attract our attention because they allow optimized neural network models to be efficiently deployed on platforms with limited energy and storage capacity. However, conventional model compression techniques prune the DNN after it is fully trained, which is very time-consuming especially when the model is trained using DRL. To overcome the limitation, we present an early phase integrated neural network weight compression system for DRL based waypoints planning. By applying pruning at an early phase, the compression of the DRL model can be realized without significant overhead in training. By tightly integrating pruning and retraining at the early phase, we achieve a higher model compression rate, reduce more memory and computing complexity, and improve the success rate compared to the original work
    corecore