2,100 research outputs found
Optimal designs for rational function regression
We consider optimal non-sequential designs for a large class of (linear and
nonlinear) regression models involving polynomials and rational functions with
heteroscedastic noise also given by a polynomial or rational weight function.
The proposed method treats D-, E-, A-, and -optimal designs in a
unified manner, and generates a polynomial whose zeros are the support points
of the optimal approximate design, generalizing a number of previously known
results of the same flavor. The method is based on a mathematical optimization
model that can incorporate various criteria of optimality and can be solved
efficiently by well established numerical optimization methods. In contrast to
previous optimization-based methods proposed for similar design problems, it
also has theoretical guarantee of its algorithmic efficiency; in fact, the
running times of all numerical examples considered in the paper are negligible.
The stability of the method is demonstrated in an example involving high degree
polynomials. After discussing linear models, applications for finding locally
optimal designs for nonlinear regression models involving rational functions
are presented, then extensions to robust regression designs, and trigonometric
regression are shown. As a corollary, an upper bound on the size of the support
set of the minimally-supported optimal designs is also found. The method is of
considerable practical importance, with the potential for instance to impact
design software development. Further study of the optimality conditions of the
main optimization model might also yield new theoretical insights.Comment: 25 pages. Previous version updated with more details in the theory
and additional example
Conic Optimization Theory: Convexification Techniques and Numerical Algorithms
Optimization is at the core of control theory and appears in several areas of
this field, such as optimal control, distributed control, system
identification, robust control, state estimation, model predictive control and
dynamic programming. The recent advances in various topics of modern
optimization have also been revamping the area of machine learning. Motivated
by the crucial role of optimization theory in the design, analysis, control and
operation of real-world systems, this tutorial paper offers a detailed overview
of some major advances in this area, namely conic optimization and its emerging
applications. First, we discuss the importance of conic optimization in
different areas. Then, we explain seminal results on the design of hierarchies
of convex relaxations for a wide range of nonconvex problems. Finally, we study
different numerical algorithms for large-scale conic optimization problems.Comment: 18 page
Adjoint-based predictor-corrector sequential convex programming for parametric nonlinear optimization
This paper proposes an algorithmic framework for solving parametric
optimization problems which we call adjoint-based predictor-corrector
sequential convex programming. After presenting the algorithm, we prove a
contraction estimate that guarantees the tracking performance of the algorithm.
Two variants of this algorithm are investigated. The first one can be used to
solve nonlinear programming problems while the second variant is aimed to treat
online parametric nonlinear programming problems. The local convergence of
these variants is proved. An application to a large-scale benchmark problem
that originates from nonlinear model predictive control of a hydro power plant
is implemented to examine the performance of the algorithms.Comment: This manuscript consists of 25 pages and 7 figure
- …