85 research outputs found
Copositivity and constrained fractional quadratic programs
Abstract We provide Completely Positive and Copositive Optimization formulations for the Constrained Fractional Quadratic Problem (CFQP) and Standard Fractional Quadratic Problem (StFQP). Based on these formulations, Semidefinite Programming (SDP) relaxations are derived for finding good lower bounds to these fractional programs, which can be used in a global optimization branch-and-bound approach. Applications of the CFQP and StFQP, related with the correction of infeasible linear systems and eigenvalue complementarity problems are also discussed
On the Influence of Multiplication Operators on the Ill-posedness of Inverse Problems: Zum Einfluss von Multiplikationsoperatoren auf die Inkorrektheit Inverser Probleme
In this thesis we deal with the degree of ill-posedness of linear operator equations in Hilbert spaces, where the operator may be decomposed into a compact linear integral operator with a well-known decay rate of singular values and a multiplication operator.
This case occurs for example for nonlinear operator equations, where the local degree of ill-posedness is investigated via the
Frechet derivative.
If the multiplier function has got zeroes, the determination of the local degree of ill-posedness is not trivial. We are going to investigate this situation, provide analytical tools as well as their limitations. By using several numerical
approaches for computing the singular values of the operator we find that the degree of ill-posedness does not change through those multiplication operators. We even provide a conjecture, verified by several numerical studies, how these multiplication operators influence the singular values of the operator equation.
Finally we analyze the influence of those multiplication operators on the opportunities of Tikhonov regularization and corresponding convergence rates. In this context we also provide a short summary on the relationship between
nonlinear problems and their linearizations.Diese Arbeit beschaeftigt sich mit dem Grad der Inkorrektheit linearer Operatorgleichungen in Hilbertraeumen, die sich als Komposition eines vollstetigen linearen Integraloperators mit bekannter Abklingrate der Singulaerwerte und eines Multiplikationsoperators darstellen lassen.
Dieser Fall tritt beispielsweise bei nichtlinearen Operatorgleichungen auf, wobei der lokale Inkorrektheitsgrad ueber die Frechetableitung bestimmt wird.
Falls die Multiplikatorfunktion Nullstellen hat, so ist die Bestimmung des lokalen Grades der Inkorrektheit nicht einfach. Moeglichkeiten und Grenzen der Analysis fuer diese Situation werden betrachtet.
Unterschiedliche numerische Ansaetze fuer die Bestimmung der Singulaerwerte liefern, dass der Grad der Inkorrektheit durch die Multiplikationsoperatoren nicht veraendert wird.
Es wird sogar ein Zusammenhang angegeben, wie Multiplikationsoperatoren die Singulaerwerte beeinflussen.
Schliesslich werden Moeglichkeiten der Tikhonov-Regularisierung unter Einfluss der Multiplikationsoperatoren untersucht. In diesem Zusammenhang wird auch eine kurze Zusammenfassung zur Beziehung von nichtlinearen Problemen und ihren Linearisierungen gegeben
Regularized interior point methods for convex programming
Interior point methods (IPMs) constitute one of the most important classes of optimization methods, due to their unparalleled robustness, as well as their generality. It is well known that a very large class of convex optimization problems can be solved by means of IPMs, in a polynomial number of iterations. As a result, IPMs are being used to solve problems arising in a plethora of fields, ranging from physics, engineering, and mathematics, to the social sciences, to name just a few. Nevertheless, there remain certain numerical issues that have not yet been addressed. More specifically, the main drawback of IPMs is that the linear algebra task involved is inherently ill-conditioned. At every iteration of the method, one has to solve a (possibly large-scale) linear system of equations (also known as the Newton system), the conditioning of which deteriorates as the IPM converges to an optimal solution. If these linear systems are of very large dimension, prohibiting the use of direct factorization, then iterative schemes may have to be employed. Such schemes are significantly affected by the inherent ill-conditioning within IPMs.
One common approach for improving the aforementioned numerical issues, is to employ regularized IPM variants. Such methods tend to be more robust and numerically stable in practice. Over the last two decades, the theory behind regularization has been significantly advanced. In particular, it is well known that regularized IPM variants can be interpreted as hybrid approaches combining IPMs with the proximal point method. However, it remained unknown whether regularized IPMs retain the polynomial complexity of their non-regularized counterparts. Furthermore, the very important issue of tuning the regularization parameters appropriately, which is also crucial in augmented Lagrangian methods, was not addressed.
In this thesis, we focus on addressing the previous open questions, as well as on creating robust implementations that solve various convex optimization problems. We discuss in detail the effect of regularization, and derive two different regularization strategies; one based on the proximal method of multipliers, and another one based on a Bregman proximal point method. The latter tends to be more efficient, while the former is more robust and has better convergence guarantees. In addition, we discuss the use of iterative linear algebra within the presented algorithms, by proposing some general purpose preconditioning strategies (used to accelerate the iterative schemes) that take advantage of the regularized nature of the systems being solved.
In Chapter 2 we present a dynamic non-diagonal regularization for IPMs. The non-diagonal aspect of this regularization is implicit, since all the off-diagonal elements of the regularization matrices are cancelled out by those elements present in the Newton system, which do not contribute important information in the computation of the Newton direction. Such a regularization, which can be interpreted as the application of a Bregman proximal point method, has multiple goals. The obvious one is to improve the spectral properties of the Newton system solved at each IPM iteration. On the other hand, the regularization matrices introduce sparsity to the aforementioned linear system, allowing for more efficient factorizations. We propose a rule for tuning the regularization dynamically based on the properties of the problem, such that sufficiently large eigenvalues of the non-regularized system are perturbed insignificantly. This alleviates the need of finding specific regularization values through experimentation, which is the most common approach in the literature. We provide perturbation bounds for the eigenvalues of the non-regularized system matrix, and then discuss the spectral properties of the regularized matrix. Finally, we demonstrate the efficiency of the method applied to solve standard small- and medium-scale linear and convex quadratic programming test problems.
In Chapter 3 we combine an IPM with the proximal method of multipliers (PMM). The resulting algorithm (IP-PMM) is interpreted as a primal-dual regularized IPM, suitable for solving linearly constrained convex quadratic programming problems. We apply few iterations of the interior point method to each sub-problem of the proximal method of multipliers. Once a satisfactory solution of the PMM sub-problem is found, we update the PMM parameters, form a new IPM neighbourhood, and repeat this process. Given this framework, we prove polynomial complexity of the algorithm, under standard assumptions. To our knowledge, this is the first polynomial complexity result for a primal-dual regularized IPM. The algorithm is guided by the use of a single penalty parameter; that of the logarithmic barrier. In other words, we show that IP-PMM inherits the polynomial complexity of IPMs, as well as the strong convexity of the PMM sub-problems. The updates of the penalty parameter are controlled by IPM, and hence are well-tuned, and do not depend on the problem solved. Furthermore, we study the behavior of the method when it is applied to an infeasible problem, and identify a necessary condition for infeasibility. The latter is used to construct an infeasibility detection mechanism. Subsequently, we provide a robust implementation of the presented algorithm and test it over a set of small to large scale linear and convex quadratic programming problems, demonstrating the benefits of using regularization in IPMs as well as the reliability of the approach.
In Chapter 4 we extend IP-PMM to the case of linear semi-definite programming (SDP) problems. In particular, we prove polynomial complexity of the algorithm, under mild assumptions, and without requiring exact computations for the Newton directions. We furthermore provide a necessary condition for lack of strong duality, which can be used as a basis for constructing detection mechanisms for identifying pathological cases within IP-PMM.
In Chapter 5 we present general-purpose preconditioners for regularized Newton systems arising within regularized interior point methods. We discuss positive definite preconditioners, suitable for iterative schemes like the conjugate gradient (CG), or the minimal residual (MINRES) method. We study the spectral properties of the preconditioned systems, and discuss the use of each presented approach, depending on the properties of the problem under consideration. All preconditioning strategies are numerically tested on various medium- to large-scale problems coming from standard test sets, as well as problems arising from partial differential equation (PDE) optimization.
In Chapter 6 we apply specialized regularized IPM variants to problems arising from portfolio optimization, machine learning, image processing, and statistics. Such problems are usually solved by specialized first-order approaches. The efficiency of the proposed regularized IPM variants is confirmed by comparing them against problem-specific state--of--the--art first-order alternatives given in the literature.
Finally, in Chapter 7 we present some conclusions as well as open questions, and possible future research directions
Recommended from our members
Learning Theory and Approximation
The main goal of this workshop – the third one of this type at the MFO – has been to blend mathematical results from statistical learning theory and approximation theory to strengthen both disciplines and use synergistic effects to work on current research questions. Learning theory aims at modeling unknown function relations and data structures from samples in an automatic manner. Approximation theory is naturally used for the advancement and closely connected to the further development of learning theory, in particular for the exploration of new useful algorithms, and for the theoretical understanding of existing methods. Conversely, the study of learning theory also gives rise to interesting theoretical problems for approximation theory such as the approximation and sparse representation of functions or the construction of rich kernel reproducing Hilbert spaces on general metric spaces. This workshop has concentrated on the following recent topics: Pitchfork bifurcation of dynamical systems arising from mathematical foundations of cell development; regularized kernel based learning in the Big Data situation; deep learning; convergence rates of learning and online learning algorithms; numerical refinement algorithms to learning; statistical robustness of regularized kernel based learning
Recommended from our members
Estimation of discriminant analysis error rate for high dimensional data
Methodologies for data reduction, modeling, and classification of grouped
response curves are explored. In particular, the thesis focuses on the analysis of
a collection of highly correlated, highly dimensional response-curve data of
spectral reflectance curves of wood surface features.
In the analysis, questions about the application of cross-validation
estimation of discriminant function error rates for data that has been previously
transformed by principal component analysis arise. Performing cross-validation
requires re-calculating the principal component transformation and discriminant
functions of the training sets, a very lengthy process. A more efficient approach
of carrying out the cross-validation calculations, plus the alternative of
estimating error rates without the re-calculation of the principal component
decomposition, are studied to address questions about the cross-validation
procedure.
If populations are assumed to have common covariance structures, the
pooled covariance matrix can be decomposed for the principal component
transformation. The leave-one-out cross-validation procedure results in a rank-one
update in the pooled covariance matrix for each observation left out.
Algorithms have been developed for calculating the updated eigenstructure
under rank-one updates and they can be applied to the orthogonal
decomposition of the pooled covariance matrix. Use of these algorithms results
in much faster computation of error rates, especially when the number of
variables is large.
The bias and variance of an estimator that performs leave-one-out cross-validation
directly on the principal component scores (without re-computation
of the principal component transformation for each observation) is also
investigated
Many Physical Design Problems are Sparse QCQPs
Physical design refers to mathematical optimization of a desired objective
(e.g. strong light--matter interactions, or complete quantum state transfer)
subject to the governing dynamical equations, such as Maxwell's or
Schrodinger's differential equations. Computing an optimal design is
challenging: generically, these problems are highly nonconvex and finding
global optima is NP hard. Here we show that for linear-differential-equation
dynamics (as in linear electromagnetism, elasticity, quantum mechanics, etc.),
the physical-design optimization problem can be transformed to a sparse-matrix,
quadratically constrained quadratic program (QCQP). Sparse QCQPs can be tackled
with convex optimization techniques (such as semidefinite programming) that
have thrived for identifying global bounds and high-performance designs in
other areas of science and engineering, but seemed inapplicable to the design
problems of wave physics. We apply our formulation to prototypical photonic
design problems, showing the possibility to compute fundamental limits for
large-area metasurfaces, as well as the identification of designs approaching
global optimality. Looking forward, our approach highlights the promise of
developing bespoke algorithms tailored to specific physical design problems.Comment: 9 pages, 4 figures, plus references and Supplementary Material
International Conference on Continuous Optimization (ICCOPT) 2019 Conference Book
The Sixth International Conference on Continuous Optimization took place on the campus of the Technical University of Berlin, August 3-8, 2019. The ICCOPT is a flagship conference of the Mathematical Optimization Society (MOS), organized every three years. ICCOPT 2019 was hosted by the Weierstrass Institute for Applied Analysis and Stochastics (WIAS) Berlin. It included a Summer School and a Conference with a series of plenary and semi-plenary talks, organized and contributed sessions, and poster sessions.
This book comprises the full conference program. It contains, in particular, the scientific program in survey style as well as with all details, and information on the social program, the venue, special meetings, and more
- …