37,090 research outputs found
Inexact Newton Methods Applied to Under-Determined Systems
Consider an under-determined system of nonlinear equations F(x)=0, F:R^m→R^n, where F is continuously differentiable and m \u3e n. This system appears in a variety of applications, including parameter-dependent systems, dynamical systems with periodic solutions, and nonlinear eigenvalue problems. Robust, efficient numerical methods are often required for the solution of this system. Newton\u27s method is an iterative scheme for solving the nonlinear system of equations F(x)=0, F:R^n→R^n. Simple to implement and theoretically sound, it is not, however, often practical in its pure form. Inexact Newton methods and globalized inexact Newton methods are computationally efficient variations of Newton\u27s method commonly used on large-scale problems. Frequently, these variations are more robust than Newton\u27s method. Trust region methods, thought of here as globalized exact Newton methods, are not as computationally efficient in the large-scale case, yet notably more robust than Newton\u27s method in practice. The normal flow method is a generalization of Newton\u27s method for solving the system F:R^m→R^n, m \u3e n. Easy to implement, this method has a simple and useful local convergence theory; however, in its pure form, it is not well suited for solving large-scale problems. This dissertation presents new methods that improve the efficiency and robustness of the normal flow method in the large-scale case. These are developed in direct analogy with inexact-Newton, globalized inexact-Newton, and trust-region methods, with particular consideration of the associated convergence theory. Included are selected problems of interest simulated in MATLAB
Recommended from our members
TENSOLVE: A software package for solving systems of nonlinear equations and nonlinear least squares problems using tensor methods
This paper describes a modular software package for solving systems of nonlinear equations and nonlinear least squares problems, using a new class of methods called tensor methods. It is intended for small to medium-sized problems, say with up to 100 equations and unknowns, in cases where it is reasonable to calculate the Jacobian matrix or approximate it by finite differences at each iteration. The software allows the user to select between a tensor method and a standard method based upon a linear model. The tensor method models F({ital x}) by a quadratic model, where the second-order term is chosen so that the model is hardly more expensive to form, store, or solve than the standard linear model. Moreover, the software provides two different global strategies, a line search and a two- dimensional trust region approach. Test results indicate that, in general, tensor methods are significantly more efficient and robust than standard methods on small and medium-sized problems in iterations and function evaluations
Progressive construction of a parametric reduced-order model for PDE-constrained optimization
An adaptive approach to using reduced-order models as surrogates in
PDE-constrained optimization is introduced that breaks the traditional
offline-online framework of model order reduction. A sequence of optimization
problems constrained by a given Reduced-Order Model (ROM) is defined with the
goal of converging to the solution of a given PDE-constrained optimization
problem. For each reduced optimization problem, the constraining ROM is trained
from sampling the High-Dimensional Model (HDM) at the solution of some of the
previous problems in the sequence. The reduced optimization problems are
equipped with a nonlinear trust-region based on a residual error indicator to
keep the optimization trajectory in a region of the parameter space where the
ROM is accurate. A technique for incorporating sensitivities into a
Reduced-Order Basis (ROB) is also presented, along with a methodology for
computing sensitivities of the reduced-order model that minimizes the distance
to the corresponding HDM sensitivity, in a suitable norm. The proposed reduced
optimization framework is applied to subsonic aerodynamic shape optimization
and shown to reduce the number of queries to the HDM by a factor of 4-5,
compared to the optimization problem solved using only the HDM, with errors in
the optimal solution far less than 0.1%
On an adaptive regularization for ill-posed nonlinear systems and its trust-region implementation
In this paper we address the stable numerical solution of nonlinear ill-posed
systems by a trust-region method. We show that an appropriate choice of the
trust-region radius gives rise to a procedure that has the potential to
approach a solution of the unperturbed system. This regularizing property is
shown theoretically and validated numerically.Comment: arXiv admin note: text overlap with arXiv:1410.278
On affine scaling inexact dogleg methods for bound-constrained nonlinear systems
Within the framework of affine scaling trust-region methods for bound constrained problems, we discuss the use of a inexact dogleg method as a tool for simultaneously handling the trust-region and the bound constraints while seeking for an approximate minimizer of the model. Focusing on bound-constrained systems of nonlinear equations, an inexact affine scaling method for large scale problems, employing the inexact dogleg procedure, is described. Global convergence results are established without any Lipschitz assumption on the Jacobian matrix, and locally fast convergence is shown under standard assumptions. Convergence analysis is performed without specifying the scaling matrix used to handle the bounds, and a rather general class of scaling matrices is allowed in actual algorithms. Numerical results showing the performance of the method are also given
- …