28,151 research outputs found

    Robust Model Predictive Control via Scenario Optimization

    Full text link
    This paper discusses a novel probabilistic approach for the design of robust model predictive control (MPC) laws for discrete-time linear systems affected by parametric uncertainty and additive disturbances. The proposed technique is based on the iterated solution, at each step, of a finite-horizon optimal control problem (FHOCP) that takes into account a suitable number of randomly extracted scenarios of uncertainty and disturbances, followed by a specific command selection rule implemented in a receding horizon fashion. The scenario FHOCP is always convex, also when the uncertain parameters and disturbance belong to non-convex sets, and irrespective of how the model uncertainty influences the system's matrices. Moreover, the computational complexity of the proposed approach does not depend on the uncertainty/disturbance dimensions, and scales quadratically with the control horizon. The main result in this paper is related to the analysis of the closed loop system under receding-horizon implementation of the scenario FHOCP, and essentially states that the devised control law guarantees constraint satisfaction at each step with some a-priori assigned probability p, while the system's state reaches the target set either asymptotically, or in finite time with probability at least p. The proposed method may be a valid alternative when other existing techniques, either deterministic or stochastic, are not directly usable due to excessive conservatism or to numerical intractability caused by lack of convexity of the robust or chance-constrained optimization problem.Comment: This manuscript is a preprint of a paper accepted for publication in the IEEE Transactions on Automatic Control, with DOI: 10.1109/TAC.2012.2203054, and is subject to IEEE copyright. The copy of record will be available at http://ieeexplore.ieee.or

    Robust stability of min-max MPC controllers for nonlinear systems with bounded uncertainties

    Get PDF
    Sixteenth International Symposium on Mathematical Theory of Networks and Systems 05/07/2004 Leuven, BélgicaThe closed loop formulation of the robust MPC has been shown to be a control technique capable of robustly stabilize uncertain nonlinear systems subject to constraints. Robust asymptotic stability of these controllers has been proved when the uncertainties are decaying. In this paper we extend the existing results to the case of uncertainties that decay with the state but do not tend to zero. This allows us to consider both plant uncertainties and external disturbances in a less conservative way. First, we provide some results on robust stability under the considered kind of uncertainties. Based on these, we prove robust stability of the min-max MPC. In the paper we show how the robust design of the local controller is translated to the min-max controller and how the persistent term of the uncertainties determines the convergence rate of the closed-loop system.Ministerio de Ciencia y Tecnología DPI-2001-2380-03-01Ministerio de Ciencia y Tecnología DPI-2002-4375-C02-0

    Computational burden reduction in Min-Max MPC

    Get PDF
    Min–max model predictive control (MMMPC) is one of the strategies used to control plants subject to bounded uncertainties. The implementation of MMMPC suffers a large computational burden due to the complex numerical optimization problem that has to be solved at every sampling time. This paper shows how to overcome this by transforming the original problem into a reduced min–max problem whose solution is much simpler. In this way, the range of processes to which MMMPC can be applied is considerably broadened. Proofs based on the properties of the cost function and simulation examples are given in the paper

    Robust Constrained Model Predictive Control using Linear Matrix Inequalities

    Get PDF
    The primary disadvantage of current design techniques for model predictive control (MPC) is their inability to deal explicitly with plant model uncertainty. In this paper, we present a new approach for robust MPC synthesis which allows explicit incorporation of the description of plant uncertainty in the problem formulation. The uncertainty is expressed both in the time domain and the frequency domain. The goal is to design, at each time step, a state-feedback control law which minimizes a "worst-case" infinite horizon objective function, subject to constraints on the control input and plant output. Using standard techniques, the problem of minimizing an upper bound on the "worst-case" objective function, subject to input and output constraints, is reduced to a convex optimization involving linear matrix inequalities (LMIs). It is shown that the feasible receding horizon state-feedback control design robustly stabilizes the set of uncertain plants under consideration. Several extensions, such as application to systems with time-delays and problems involving constant set-point tracking, trajectory tracking and disturbance rejection, which follow naturally from our formulation, are discussed. The controller design procedure is illustrated with two examples. Finally, conclusions are presented

    From Uncertainty Data to Robust Policies for Temporal Logic Planning

    Full text link
    We consider the problem of synthesizing robust disturbance feedback policies for systems performing complex tasks. We formulate the tasks as linear temporal logic specifications and encode them into an optimization framework via mixed-integer constraints. Both the system dynamics and the specifications are known but affected by uncertainty. The distribution of the uncertainty is unknown, however realizations can be obtained. We introduce a data-driven approach where the constraints are fulfilled for a set of realizations and provide probabilistic generalization guarantees as a function of the number of considered realizations. We use separate chance constraints for the satisfaction of the specification and operational constraints. This allows us to quantify their violation probabilities independently. We compute disturbance feedback policies as solutions of mixed-integer linear or quadratic optimization problems. By using feedback we can exploit information of past realizations and provide feasibility for a wider range of situations compared to static input sequences. We demonstrate the proposed method on two robust motion-planning case studies for autonomous driving
    corecore