1,090 research outputs found
Bounding stationary averages of polynomial diffusions via semidefinite programming
We introduce an algorithm based on semidefinite programming that yields
increasing (resp. decreasing) sequences of lower (resp. upper) bounds on
polynomial stationary averages of diffusions with polynomial drift vector and
diffusion coefficients. The bounds are obtained by optimising an objective,
determined by the stationary average of interest, over the set of real vectors
defined by certain linear equalities and semidefinite inequalities which are
satisfied by the moments of any stationary measure of the diffusion. We
exemplify the use of the approach through several applications: a Bayesian
inference problem; the computation of Lyapunov exponents of linear ordinary
differential equations perturbed by multiplicative white noise; and a
reliability problem from structural mechanics. Additionally, we prove that the
bounds converge to the infimum and supremum of the set of stationary averages
for certain SDEs associated with the computation of the Lyapunov exponents, and
we provide numerical evidence of convergence in more general settings
Controller Synthesis for Discrete-Time Polynomial Systems via Occupation Measures
In this paper, we design nonlinear state feedback controllers for
discrete-time polynomial dynamical systems via the occupation measure approach.
We propose the discrete-time controlled Liouville equation, and use it to
formulate the controller synthesis problem as an infinite-dimensional linear
programming problem on measures, which is then relaxed as finite-dimensional
semidefinite programming problems on moments of measures and their duals on
sums-of-squares polynomials. Nonlinear controllers can be extracted from the
solutions to the relaxed problems. The advantage of the occupation measure
approach is that we solve convex problems instead of generally non-convex
problems, and the computational complexity is polynomial in the state and input
dimensions, and hence the approach is more scalable. In addition, we show that
the approach can be applied to over-approximating the backward reachable set of
discrete-time autonomous polynomial systems and the controllable set of
discrete-time polynomial systems under known state feedback control laws. We
illustrate our approach on several dynamical systems
Linearly Solvable Stochastic Control Lyapunov Functions
This paper presents a new method for synthesizing stochastic control Lyapunov
functions for a class of nonlinear stochastic control systems. The technique
relies on a transformation of the classical nonlinear Hamilton-Jacobi-Bellman
partial differential equation to a linear partial differential equation for a
class of problems with a particular constraint on the stochastic forcing. This
linear partial differential equation can then be relaxed to a linear
differential inclusion, allowing for relaxed solutions to be generated using
sum of squares programming. The resulting relaxed solutions are in fact
viscosity super/subsolutions, and by the maximum principle are pointwise upper
and lower bounds to the underlying value function, even for coarse polynomial
approximations. Furthermore, the pointwise upper bound is shown to be a
stochastic control Lyapunov function, yielding a method for generating
nonlinear controllers with pointwise bounded distance from the optimal cost
when using the optimal controller. These approximate solutions may be computed
with non-increasing error via a hierarchy of semidefinite optimization
problems. Finally, this paper develops a-priori bounds on trajectory
suboptimality when using these approximate value functions, as well as
demonstrates that these methods, and bounds, can be applied to a more general
class of nonlinear systems not obeying the constraint on stochastic forcing.
Simulated examples illustrate the methodology.Comment: Published in SIAM Journal of Control and Optimizatio
A New Distribution-Free Concept for Representing, Comparing, and Propagating Uncertainty in Dynamical Systems with Kernel Probabilistic Programming
This work presents the concept of kernel mean embedding and kernel
probabilistic programming in the context of stochastic systems. We propose
formulations to represent, compare, and propagate uncertainties for fairly
general stochastic dynamics in a distribution-free manner. The new tools enjoy
sound theory rooted in functional analysis and wide applicability as
demonstrated in distinct numerical examples. The implication of this new
concept is a new mode of thinking about the statistical nature of uncertainty
in dynamical systems
Some Applications of Polynomial Optimization in Operations Research and Real-Time Decision Making
We demonstrate applications of algebraic techniques that optimize and certify
polynomial inequalities to problems of interest in the operations research and
transportation engineering communities. Three problems are considered: (i)
wireless coverage of targeted geographical regions with guaranteed signal
quality and minimum transmission power, (ii) computing real-time certificates
of collision avoidance for a simple model of an unmanned vehicle (UV)
navigating through a cluttered environment, and (iii) designing a nonlinear
hovering controller for a quadrotor UV, which has recently been used for load
transportation. On our smaller-scale applications, we apply the sum of squares
(SOS) relaxation and solve the underlying problems with semidefinite
programming. On the larger-scale or real-time applications, we use our recently
introduced "SDSOS Optimization" techniques which result in second order cone
programs. To the best of our knowledge, this is the first study of real-time
applications of sum of squares techniques in optimization and control. No
knowledge in dynamics and control is assumed from the reader
Stochastic MPC Design for a Two-Component Granulation Process
We address the issue of control of a stochastic two-component granulation
process in pharmaceutical applications through using Stochastic Model
Predictive Control (SMPC) and model reduction to obtain the desired particle
distribution. We first use the method of moments to reduce the governing
integro-differential equation down to a nonlinear ordinary differential
equation (ODE). This reduced-order model is employed in the SMPC formulation.
The probabilistic constraints in this formulation keep the variance of
particles' drug concentration in an admissible range. To solve the resulting
stochastic optimization problem, we first employ polynomial chaos expansion to
obtain the Probability Distribution Function (PDF) of the future state
variables using the uncertain variables' distributions. As a result, the
original stochastic optimization problem for a particulate system is converted
to a deterministic dynamic optimization. This approximation lessens the
computation burden of the controller and makes its real time application
possible.Comment: American control Conference, May, 201
Mini-Workshop: Applied Koopmanism
Koopman and Perron–Frobenius operators are linear operators that encapsulate dynamics of nonlinear dynamical systems without loss of information. This is accomplished by embedding the dynamics into a larger infinite-dimensional space where the focus of study is shifted from trajectory curves to measurement functions evaluated along trajectories and densities of trajectories evolving in time. Operator-theoretic approach to dynamics shares many features with an optimization technique: the Lasserre moment–sums-of-squares (SOS) hierarchies, which was developed for numerically solving non-convex optimization problems with semialgebraic data. This technique embeds the optimization problem into a larger primal semidefinite programming (SDP) problem consisting of measure optimization over the set of globally optimal solutions, where measures are manipulated through their truncated moment sequences. The dual SDP problem uses SOS representations to certify bounds on the global optimum. This workshop highlighted the common threads between the operator-theoretic dynamical systems and moment–SOS hierarchies in optimization and explored the future directions where the synergy of the two techniques could yield results in fluid dynamics, control theory, optimization, and spectral theory
- …