1,272 research outputs found

    Efficient Algorithms for Optimal Control of Quantum Dynamics: The "Krotov'' Method unencumbered

    Full text link
    Efficient algorithms for the discovery of optimal control designs for coherent control of quantum processes are of fundamental importance. One important class of algorithms are sequential update algorithms generally attributed to Krotov. Although widely and often successfully used, the associated theory is often involved and leaves many crucial questions unanswered, from the monotonicity and convergence of the algorithm to discretization effects, leading to the introduction of ad-hoc penalty terms and suboptimal update schemes detrimental to the performance of the algorithm. We present a general framework for sequential update algorithms including specific prescriptions for efficient update rules with inexpensive dynamic search length control, taking into account discretization effects and eliminating the need for ad-hoc penalty terms. The latter, while necessary to regularize the problem in the limit of infinite time resolution, i.e., the continuum limit, are shown to be undesirable and unnecessary in the practically relevant case of finite time resolution. Numerical examples show that the ideas underlying many of these results extend even beyond what can be rigorously proved.Comment: 19 pages, many figure

    Asymptotic expansions for interior solutions of semilinear elliptic problems

    Get PDF
    International audienceIn this work we consider the optimal control problem of a semilinear elliptic PDE with a Dirichlet boundary condition, where the control variable is distributed over the domain and is constrained to be nonnegative. The approach is to consider an associated parametrized family of penalized problems, whose solutions define a central path converging to the solution of the original problem. Our aim is to obtain an asymptotic expansion for the solutions of the penalized problems around the solution of the original problem. This approach allows us to obtain some specific error bounds in various norms and for a general class of barrier functions. In this manner, we generalize the results of the previous work which were obtained in the ODE framework

    A second derivative SQP method: theoretical issues

    Get PDF
    Sequential quadratic programming (SQP) methods form a class of highly efficient algorithms for solving nonlinearly constrained optimization problems. Although second derivative information may often be calculated, there is little practical theory that justifies exact-Hessian SQP methods. In particular, the resulting quadratic programming (QP) subproblems are often nonconvex, and thus finding their global solutions may be computationally nonviable. This paper presents a second-derivative SQP method based on quadratic subproblems that are either convex, and thus may be solved efficiently, or need not be solved globally. Additionally, an explicit descent-constraint is imposed on certain QP subproblems, which “guides” the iterates through areas in which nonconvexity is a concern. Global convergence of the resulting algorithm is established

    On the Effectiveness of Richardson Extrapolation in Machine Learning

    Get PDF
    Richardson extrapolation is a classical technique from numerical analysis that can improve the approximation error of an estimation method by combining linearly several estimates obtained from different values of one of its hyperparameters, without the need to know in details the inner structure of the original estimation method. The main goal of this paper is to study when Richardson extrapolation can be used within machine learning, beyond the existing applications to step-size adaptations in stochastic gradient descent. We identify two situations where Richardson interpolation can be useful: (1) when the hyperparameter is the number of iterations of an existing iterative optimization algorithm, with applications to averaged gradient descent and Frank-Wolfe algorithms (where we obtain asymptotically rates of O(1/k2)O(1/k^2) on polytopes, where kk is the number of iterations), and (2) when it is a regularization parameter, with applications to Nesterov smoothing techniques for minimizing non-smooth functions (where we obtain asymptotically rates close to O(1/k2)O(1/k^2) for non-smooth functions), and ridge regression. In all these cases, we show that extrapolation techniques come with no significant loss in performance, but with sometimes strong gains, and we provide theoretical justifications based on asymptotic developments for such gains, as well as empirical illustrations on classical problems from machine learning

    Model Uncertainty, Recalibration, and the Emergence of Delta-Vega Hedging

    Full text link
    We study option pricing and hedging with uncertainty about a Black-Scholes reference model which is dynamically recalibrated to the market price of a liquidly traded vanilla option. For dynamic trading in the underlying asset and this vanilla option, delta-vega hedging is asymptotically optimal in the limit for small uncertainty aversion. The corresponding indifference price corrections are determined by the disparity between the vegas, gammas, vannas, and volgas of the non-traded and the liquidly traded options.Comment: 44 pages; forthcoming in 'Finance and Stochastics

    Error estimates for the logarithmic barrier method in stochastic linear quadratic optimal control problems

    Get PDF
    International audienceWe consider a linear quadratic stochastic optimal control problem whith non-negativity control constraints. The latter are penalized with the classical logarithmic barrier. Using a duality argument and the stochastic minimum principle, we provide an error estimate for the solution of the penalized problem which is the natural extension of the well known estimate in the deterministic framework

    Distributed Receding Horizon Control with Application to Multi-Vehicle Formation Stabilization

    Get PDF
    We consider the control of interacting subsystems whose dynamics and constraints are uncoupled, but whose state vectors are coupled non-separably in a single centralized cost function of a finite horizon optimal control problem. For a given centralized cost structure, we generate distributed optimal control problems for each subsystem and establish that the distributed receding horizon implementation is asymptotically stabilizing. The communication requirements between subsystems with coupling in the cost function are that each subsystem obtain the previous optimal control trajectory of those subsystems at each receding horizon update. The key requirements for stability are that each distributed optimal control not deviate too far from the previous optimal control, and that the receding horizon updates happen sufficiently fast. The theory is applied in simulation for stabilization of a formation of vehicles

    A Bayesian information criterion for singular models

    Full text link
    We consider approximate Bayesian model choice for model selection problems that involve models whose Fisher-information matrices may fail to be invertible along other competing submodels. Such singular models do not obey the regularity conditions underlying the derivation of Schwarz's Bayesian information criterion (BIC) and the penalty structure in BIC generally does not reflect the frequentist large-sample behavior of their marginal likelihood. While large-sample theory for the marginal likelihood of singular models has been developed recently, the resulting approximations depend on the true parameter value and lead to a paradox of circular reasoning. Guided by examples such as determining the number of components of mixture models, the number of factors in latent factor models or the rank in reduced-rank regression, we propose a resolution to this paradox and give a practical extension of BIC for singular model selection problems
    • …
    corecore