11,581 research outputs found
Singularly perturbed forward-backward stochastic differential equations: application to the optimal control of bilinear systems
We study linear-quadratic stochastic optimal control problems with bilinear
state dependence for which the underlying stochastic differential equation
(SDE) consists of slow and fast degrees of freedom. We show that, in the same
way in which the underlying dynamics can be well approximated by a reduced
order effective dynamics in the time scale limit (using classical
homogenziation results), the associated optimal expected cost converges in the
time scale limit to an effective optimal cost. This entails that we can well
approximate the stochastic optimal control for the whole system by the reduced
order stochastic optimal control, which is clearly easier to solve because of
lower dimensionality. The approach uses an equivalent formulation of the
Hamilton-Jacobi-Bellman (HJB) equation, in terms of forward-backward SDEs
(FBSDEs). We exploit the efficient solvability of FBSDEs via a least squares
Monte Carlo algorithm and show its applicability by a suitable numerical
example
Numerical computation of rare events via large deviation theory
An overview of rare events algorithms based on large deviation theory (LDT)
is presented. It covers a range of numerical schemes to compute the large
deviation minimizer in various setups, and discusses best practices, common
pitfalls, and implementation trade-offs. Generalizations, extensions, and
improvements of the minimum action methods are proposed. These algorithms are
tested on example problems which illustrate several common difficulties which
arise e.g. when the forcing is degenerate or multiplicative, or the systems are
infinite-dimensional. Generalizations to processes driven by non-Gaussian
noises or random initial data and parameters are also discussed, along with the
connection between the LDT-based approach reviewed here and other methods, such
as stochastic field theory and optimal control. Finally, the integration of
this approach in importance sampling methods using e.g. genealogical algorithms
is explored
Particle filtering in high-dimensional chaotic systems
We present an efficient particle filtering algorithm for multiscale systems,
that is adapted for simple atmospheric dynamics models which are inherently
chaotic. Particle filters represent the posterior conditional distribution of
the state variables by a collection of particles, which evolves and adapts
recursively as new information becomes available. The difference between the
estimated state and the true state of the system constitutes the error in
specifying or forecasting the state, which is amplified in chaotic systems that
have a number of positive Lyapunov exponents. The purpose of the present paper
is to show that the homogenization method developed in Imkeller et al. (2011),
which is applicable to high dimensional multi-scale filtering problems, along
with important sampling and control methods can be used as a basic and flexible
tool for the construction of the proposal density inherent in particle
filtering. Finally, we apply the general homogenized particle filtering
algorithm developed here to the Lorenz'96 atmospheric model that mimics
mid-latitude atmospheric dynamics with microscopic convective processes.Comment: 28 pages, 12 figure
An Optimal Control Derivation of Nonlinear Smoothing Equations
The purpose of this paper is to review and highlight some connections between
the problem of nonlinear smoothing and optimal control of the Liouville
equation. The latter has been an active area of recent research interest owing
to work in mean-field games and optimal transportation theory. The nonlinear
smoothing problem is considered here for continuous-time Markov processes. The
observation process is modeled as a nonlinear function of a hidden state with
an additive Gaussian measurement noise. A variational formulation is described
based upon the relative entropy formula introduced by Newton and Mitter. The
resulting optimal control problem is formulated on the space of probability
distributions. The Hamilton's equation of the optimal control are related to
the Zakai equation of nonlinear smoothing via the log transformation. The
overall procedure is shown to generalize the classical Mortensen's minimum
energy estimator for the linear Gaussian problem.Comment: 7 pages, 0 figures, under peer reviewin
- …