1,047 research outputs found

    Numerical iterative methods for nonlinear problems.

    Get PDF
    The primary focus of research in this thesis is to address the construction of iterative methods for nonlinear problems coming from different disciplines. The present manuscript sheds light on the development of iterative schemes for scalar nonlinear equations, for computing the generalized inverse of a matrix, for general classes of systems of nonlinear equations and specific systems of nonlinear equations associated with ordinary and partial differential equations. Our treatment of the considered iterative schemes consists of two parts: in the first called the ’construction part’ we define the solution method; in the second part we establish the proof of local convergence and we derive convergence-order, by using symbolic algebra tools. The quantitative measure in terms of floating-point operations and the quality of the computed solution, when real nonlinear problems are considered, provide the efficiency comparison among the proposed and the existing iterative schemes. In the case of systems of nonlinear equations, the multi-step extensions are formed in such a way that very economical iterative methods are provided, from a computational viewpoint. Especially in the multi-step versions of an iterative method for systems of nonlinear equations, the Jacobians inverses are avoided which make the iterative process computationally very fast. When considering special systems of nonlinear equations associated with ordinary and partial differential equations, we can use higher-order Frechet derivatives thanks to the special type of nonlinearity: from a computational viewpoint such an approach has to be avoided in the case of general systems of nonlinear equations due to the high computational cost. Aside from nonlinear equations, an efficient matrix iteration method is developed and implemented for the calculation of weighted Moore-Penrose inverse. Finally, a variety of nonlinear problems have been numerically tested in order to show the correctness and the computational efficiency of our developed iterative algorithms

    Convergence Rate Improvement of Richardson and Newton-Schulz Iterations

    Get PDF
    Fast convergent, accurate, computationally efficient, parallelizable, and robust matrix inversion and parameter estimation algorithms are required in many time-critical and accuracy-critical applications such as system identification, signal and image processing, network and big data analysis, machine learning and in many others. This paper introduces new composite power series expansion with optionally chosen rates (which can be calculated simultaneously on parallel units with different computational capacities) for further convergence rate improvement of high order Newton-Schulz iteration. New expansion was integrated into the Richardson iteration and resulted in significant convergence rate improvement. The improvement is quantified via explicit transient models for estimation errors and by simulations. In addition, the recursive and computationally efficient version of the combination of Richardson iteration and Newton-Schulz iteration with composite expansion is developed for simultaneous calculations. Moreover, unified factorization is developed in this paper in the form of tool-kit for power series expansion, which results in a new family of computationally efficient Newton-Schulz algorithms

    A quasi-Newton proximal splitting method

    Get PDF
    A new result in convex analysis on the calculation of proximity operators in certain scaled norms is derived. We describe efficient implementations of the proximity calculation for a useful class of functions; the implementations exploit the piece-wise linear nature of the dual problem. The second part of the paper applies the previous result to acceleration of convex minimization problems, and leads to an elegant quasi-Newton method. The optimization method compares favorably against state-of-the-art alternatives. The algorithm has extensive applications including signal processing, sparse recovery and machine learning and classification
    • …
    corecore