391 research outputs found

    Robust Model Predictive Control for Linear Parameter Varying Systems along with Exploration of its Application in Medical Mobile Robots

    Get PDF
    This thesis seeks to develop a robust model predictive controller (MPC) for Linear Parameter Varying (LPV) systems. LPV models based on input-output display are employed. We aim to improve robust MPC methods for LPV systems with an input-output display. This improvement will be examined from two perspectives. First, the system must be stable in conditions of uncertainty (in signal scheduling or due to disturbance) and perform well in both tracking and regulation problems. Secondly, the proposed method should be practical, i.e., it should have a reasonable computational load and not be conservative. Firstly, an interpolation approach is utilized to minimize the conservativeness of the MPC. The controller is calculated as a linear combination of a set of offline predefined control laws. The coefficients of these offline controllers are derived from a real-time optimization problem. The control gains are determined to ensure stability and increase the terminal set. Secondly, in order to test the system's robustness to external disturbances, a free control move was added to the control law. Also, a Recurrent Neural Network (RNN) algorithm is applied for online optimization, showing that this optimization method has better speed and accuracy than traditional algorithms. The proposed controller was compared with two methods (robust MPC and MPC with LPV model based on input-output) in reference tracking and disturbance rejection scenarios. It was shown that the proposed method works well in both parts. However, two other methods could not deal with the disturbance. Thirdly, a support vector machine was introduced to identify the input-output LPV model to estimate the output. The estimated model was compared with the actual nonlinear system outputs, and the identification was shown to be effective. As a consequence, the controller can accurately follow the reference. Finally, an interpolation-based MPC with free control moves is implemented for a wheeled mobile robot in a hospital setting, where an RNN solves the online optimization problem. The controller was compared with a robust MPC and MPC-LPV in reference tracking, disturbance rejection, online computational load, and region of attraction. The results indicate that our proposed method surpasses and can navigate quickly and reliably while avoiding obstacles

    Adaptive and learning-based formation control of swarm robots

    Get PDF
    Autonomous aerial and wheeled mobile robots play a major role in tasks such as search and rescue, transportation, monitoring, and inspection. However, these operations are faced with a few open challenges including robust autonomy, and adaptive coordination based on the environment and operating conditions, particularly in swarm robots with limited communication and perception capabilities. Furthermore, the computational complexity increases exponentially with the number of robots in the swarm. This thesis examines two different aspects of the formation control problem. On the one hand, we investigate how formation could be performed by swarm robots with limited communication and perception (e.g., Crazyflie nano quadrotor). On the other hand, we explore human-swarm interaction (HSI) and different shared-control mechanisms between human and swarm robots (e.g., BristleBot) for artistic creation. In particular, we combine bio-inspired (i.e., flocking, foraging) techniques with learning-based control strategies (using artificial neural networks) for adaptive control of multi- robots. We first review how learning-based control and networked dynamical systems can be used to assign distributed and decentralized policies to individual robots such that the desired formation emerges from their collective behavior. We proceed by presenting a novel flocking control for UAV swarm using deep reinforcement learning. We formulate the flocking formation problem as a partially observable Markov decision process (POMDP), and consider a leader-follower configuration, where consensus among all UAVs is used to train a shared control policy, and each UAV performs actions based on the local information it collects. In addition, to avoid collision among UAVs and guarantee flocking and navigation, a reward function is added with the global flocking maintenance, mutual reward, and a collision penalty. We adapt deep deterministic policy gradient (DDPG) with centralized training and decentralized execution to obtain the flocking control policy using actor-critic networks and a global state space matrix. In the context of swarm robotics in arts, we investigate how the formation paradigm can serve as an interaction modality for artists to aesthetically utilize swarms. In particular, we explore particle swarm optimization (PSO) and random walk to control the communication between a team of robots with swarming behavior for musical creation

    Design, testing and validation of model predictive control for an unmanned ground vehicle

    Full text link
    The rapid increase in designing, manufacturing, and using autonomous robots has attracted numerous researchers and industries in recent decades. The logical motivation behind this interest is the wide range of applications. For instance, perimeter surveillance, search and rescue missions, agriculture, and construction. In this thesis, motion planning and control based on model predictive control (MPC) for unmanned ground vehicles (UGVs) is tackled. In addition, different variants of MPC are designed, analysed, and implemented for such non-holonomic systems. It is imperative to focus on the ability of MPC to handle constraints as one of the motivations. Furthermore, the proliferation of computer processing enables these systems to work in a real-time scenario. The controller's responsibility is to guarantee an accurate trajectory tracking process to deal with other specifications usually not considered or solved by the planner. However, the separation between planner and controller is not necessarily defined uniquely, even though it can be a hybrid process, as seen in part of this thesis. Firstly, a robust MPC is designed and implemented for a small-scale autonomous bulldozer in the presence of uncertainties, which uses an optimal control action and a feed-forward controller to suppress these uncertainties. More precisely, a linearised variant of MPC is deployed to solve the trajectory tracking problem of the vehicle. Afterwards, a nonlinear MPC is designed and implemented to solve the path-following problem of the UGV for masonry in a construction context, where longitudinal velocity and yaw rate are employed as control inputs to the platform. For both the control techniques, several experiments are performed to validate the robustness and accuracy of the proposed scheme. Those experiments are performed under realistic localisation accuracy, provided by a typical localiser. Most conspicuously, a novel proximal planning and control strategy is implemented in the presence of skid-slip and dynamic and static collision avoidance for the posture control and tracking control problems. The ability to operate in moving objects is critical for UGVs to function well. The approach offers specific planning capabilities, able to deal at high frequency with context characteristics, which the higher-level planner may not well solve. Those context characteristics are related to dynamic objects and other terrain details detected by the platform's onboard perception capabilities. In the control context, proximal and interior-point optimisation methods are used for MPC. Relevant attention is given to the processing time required by the MPC process to obtain the control actions at each actual control time. This concern is due to the need to optimise each control action, which must be calculated and applied in real-time. Because the length of a prediction horizon is critical in practical applications, it is worth looking into in further detail. In another study, the accuracies of robust and nonlinear model predictive controllers are compared. Finally, a hybrid controller is proposed and implemented. This approach exploits the availability of a simplified cost-to-go function (which is provided by a higher-level planner); thus, the hybrid approach fuses, in real-time, the nominal CTG function (nominal terrain map) with the rest of the critical constraints, which the planner usually ignores. The conducted research fills necessary gaps in the application areas of MPC and UGVs. Both theoretical and practical contributions have been made in this thesis. Moreover, extensive simulations and experiments are performed to test and verify the working of MPC with a reasonable processing capability of the onboard process

    Adaptive dynamic programming with eligibility traces and complexity reduction of high-dimensional systems

    Get PDF
    This dissertation investigates the application of a variety of computational intelligence techniques, particularly clustering and adaptive dynamic programming (ADP) designs especially heuristic dynamic programming (HDP) and dual heuristic programming (DHP). Moreover, a one-step temporal-difference (TD(0)) and n-step TD (TD(λ)) with their gradients are utilized as learning algorithms to train and online-adapt the families of ADP. The dissertation is organized into seven papers. The first paper demonstrates the robustness of model order reduction (MOR) for simulating complex dynamical systems. Agglomerative hierarchical clustering based on performance evaluation is introduced for MOR. This method computes the reduced order denominator of the transfer function by clustering system poles in a hierarchical dendrogram. Several numerical examples of reducing techniques are taken from the literature to compare with our work. In the second paper, a HDP is combined with the Dyna algorithm for path planning. The third paper uses DHP with an eligibility trace parameter (λ) to track a reference trajectory under uncertainties for a nonholonomic mobile robot by using a first-order Sugeno fuzzy neural network structure for the critic and actor networks. In the fourth and fifth papers, a stability analysis for a model-free action-dependent HDP(λ) is demonstrated with batch- and online-implementation learning, respectively. The sixth work combines two different gradient prediction levels of critic networks. In this work, we provide a convergence proofs. The seventh paper develops a two-hybrid recurrent fuzzy neural network structures for both critic and actor networks. They use a novel n-step gradient temporal-difference (gradient of TD(λ)) of an advanced ADP algorithm called value-gradient learning (VGL(λ)), and convergence proofs are given. Furthermore, the seventh paper is the first to combine the single network adaptive critic with VGL(λ). --Abstract, page iv

    A Scalable Safety Critical Control Framework for Nonlinear Systems

    Get PDF
    There are two main approaches to safety-critical control. The first one relies on computation of control invariant sets and is presented in the first part of this work. The second approach draws from the topic of optimal control and relies on the ability to realize Model-Predictive-Controllers online to guarantee the safety of a system. In the second approach, safety is ensured at a planning stage by solving the control problem subject for some explicitly defined constraints on the state and control input. Both approaches have distinct advantages but also major drawbacks that hinder their practical effectiveness, namely scalability for the first one and computational complexity for the second. We therefore present an approach that draws from the advantages of both approaches to deliver efficient and scalable methods of ensuring safety for nonlinear dynamical systems. In particular, we show that identifying a backup control law that stabilizes the system is in fact sufficient to exploit some of the set-invariance conditions presented in the first part of this work. Indeed, one only needs to be able to numerically integrate the closed-loop dynamics of the system over a finite horizon under this backup law to compute all the information necessary for evaluating the regulation map and enforcing safety. The effect of relaxing the stabilization requirements of the backup law is also studied, and weaker but more practical safety guarantees are brought forward. We then explore the relationship between the optimality of the backup law and how conservative the resulting safety filter is. Finally, methods of selecting a safe input with varying levels of trade-off between conservatism and computational complexity are proposed and illustrated on multiple robotic systems, namely: a two-wheeled inverted pendulum (Segway), an industrial manipulator, a quadrotor, and a lower body exoskeleton

    Reachability-based Trajectory Design

    Full text link
    Autonomous mobile robots have the potential to increase the availability and accessibility of goods and services throughout society. However, to enable public trust in such systems, it is critical to certify that they are safe. This requires formally specifying safety, and designing motion planning methods that can guarantee safe operation (note, this work is only concerned with planning, not perception). The typical paradigm to attempt to ensure safety is receding-horizon planning, wherein a robot creates a short plan, then executes it while creating its next short plan in an iterative fashion, allowing a robot to incorporate new sensor information over time. However, this requires a robot to plan in real time. Therefore, the key challenge in making safety guarantees lies in balancing performance (how quickly a robot can plan) and conservatism (how cautiously a robot behaves). Existing methods suffer from a tradeoff between performance and conservatism, which is rooted in the choice of model used describe a robot; accuracy typically comes at the price of computation speed. To address this challenge, this dissertation proposes Reachability-based Trajectory Design (RTD), which performs real-time, receding-horizon planning with a simplified planning model, and ensures safety by describing the model error using a reachable set of the robot. RTD begins with the offline design of a continuum of parameterized trajectories for the plan- ning model; each trajectory ends with a fail-safe maneuver such as braking to a stop. RTD then computes the robot’s Forward Reachable Set (FRS), which contains all points in workspace reach- able by the robot for each parameterized trajectory. Importantly, the FRS also contains the error model, since a robot can typically never track planned trajectories perfectly. Online (at runtime), the robot intersects the FRS with sensed obstacles to provably determine which trajectory plans could cause collisions. Then, the robot performs trajectory optimization over the remaining safe trajectories. If no new safe plan can be found, the robot can execute its previously-found fail-safe maneuver, enabling perpetual safety. This dissertation begins by presenting RTD as a theoretical framework, then presents three representations of a robot’s FRS, using (1) sums-of-squares (SOS) polynomial programming, (2) zonotopes (a special type of convex polytope), and (3) rotatotopes (a generalization of zonotopes that enable representing a robot’s swept volume). To enable real-time planning, this work also de- velops an obstacle representation that enables provable safety while treating obstacles as discrete, finite sets of points. The practicality of RTD is demonstrated on four different wheeled robots (using the SOS FRS), two quadrotor aerial robots (using the zonotope FRS), and one manipulator robot (using the rotatotope FRS). Over thousands of simulations and dozens of hardware trials, RTD performs safe, real-time planning in arbitrary and challenging environments. In summary, this dissertation proposes RTD as a general purpose, practical framework for provably safe, real-time robot motion planning.PHDMechanical EngineeringUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/162884/1/skousik_1.pd

    Learning of Closed-Loop Motion Control

    Get PDF
    Learning motion control as a unified process of designing the reference trajectory and the controller is one of the most challenging problems in robotics. The complexity of the problem prevents most of the existing optimization algorithms from giving satisfactory results. While model-based algorithms like iterative linear-quadratic-Gaussian (iLQG) can be used to design a suitable controller for the motion control, their performance is strongly limited by the model accuracy. An inaccurate model may lead to degraded performance of the controller on the physical system. Although using machine learning approaches to learn the motion control on real systems have been proven to be effective, their performance depends on good initialization. To address these issues, this paper introduces a two-step algorithm which combines the proven performance of a model-based controller with a model-free method for compensating for model inaccuracy. The first step optimizes the problem using iLQG. Then, in the second step this controller is used to initialize the policy for our PI2^2-01 reinforcement learning algorithm. This algorithm is a derivation of the PI2^2 algorithm enabling more stable and faster convergence. The performance of this method is demonstrated both in simulation and experimental results
    corecore