413 research outputs found
Randomized Orthogonal Projection Methods for Krylov Subspace Solvers
Randomized orthogonal projection methods (ROPMs) can be used to speed up the
computation of Krylov subspace methods in various contexts. Through a
theoretical and numerical investigation, we establish that these methods
produce quasi-optimal approximations over the Krylov subspace. Our numerical
experiments outline the convergence of ROPMs for all matrices in our test set,
with occasional spikes, but overall with a convergence rate similar to that of
standard OPMs
Recommended from our members
Inexact and Nonlinear Extensions of the FEAST Eigenvalue Algorithm
Eigenvalue problems are a basic element of linear algebra that have a wide variety of applications. Common examples include determining the stability of dynamical systems, performing dimensionality reduction on large data sets, and predicting the physical properties of nanoscopic objects. Many applications require solving large dimensional eigenvalue problems, which can be very challenging when the required number of eigenvalues and eigenvectors is also large. The FEAST algorithm is a method of solving eigenvalue problems that allows one to calculate large numbers of eigenvalue/eigenvector pairs by using contour integration in the complex plane to divide the large number of desired pairs into many small groups; these small groups of eigenvalue/eigenvector pairs may then be simultaneously calculated independently of each other. This makes it possible to quickly solve eigenvalue problems that might otherwise be very difficult to solve efficiently.
The standard FEAST algorithm can only be used to solve eigenvalue problems that are linear, and whose matrices are small enough to be factorized efficiently (thus allowing linear systems of equations to be solved exactly). This limits the size and the scope of the problems to which the FEAST algorithm may be applied. This dissertation describes extensions of the standard FEAST algorithm that allow it to efficiently solve nonlinear eigenvalue problems, and eigenvalue problems whose matrices are large enough that linear systems of equations can only be solved inexactly
Analytic study on linear systems of distributed order fractional differential equations
In this paper we introduce the distributed order fractional differential equations (DOFDE) with respect to the nonnegative density function. We generalize the inertia and characteristics polynomial concepts of pair with respect to the nonnegative density function. We also give generalization of the invariant factors of a matrix and some inertia theorems for analyzing the stability of the DOFDE systems
Stability Analysis of Distributed Order Fractional Differential Equations
We analyze the stability of three classes of distributed order fractional differential equations (DOFDEs) with respect to the nonnegative density function. In this sense, we discover a robust stability condition for these systems based on characteristic function and new inertia concept of a matrix with respect to the density function. Moreover, we check the stability of a distributed order fractional WINDMI system to illustrate the validity of proposed procedure
Inner-outer Iterative Methods for Eigenvalue Problems - Convergence and Preconditioning
Many methods for computing eigenvalues of a large sparse matrix involve shift-invert transformations which require the solution of a shifted linear system at each step. This thesis deals with shift-invert iterative techniques for solving eigenvalue problems where the arising linear systems are solved inexactly using a second iterative technique. This approach leads to an inner-outer type algorithm. We provide convergence results for the outer iterative eigenvalue computation as well as techniques for efficient inner solves. In particular eigenvalue computations using inexact inverse iteration, the Jacobi-Davidson method without subspace expansion and the shift-invert Arnoldi method as a subspace method are investigated in detail. A general convergence result for inexact inverse iteration for the non-Hermitian generalised eigenvalue problem is given, using only minimal assumptions. This convergence result is obtained in two different ways; on the one hand, we use an equivalence result between inexact inverse iteration applied to the generalised eigenproblem and modified Newton's method; on the other hand, a splitting method is used which generalises the idea of orthogonal decomposition. Both approaches also include an analysis for the convergence theory of a version of inexact Jacobi-Davidson method, where equivalences between Newton's method, inverse iteration and the Jacobi-Davidson method are exploited. To improve the efficiency of the inner iterative solves we introduce a new tuning strategy which can be applied to any standard preconditioner. We give a detailed analysis on this new preconditioning idea and show how the number of iterations for the inner iterative method and hence the total number of iterations can be reduced significantly by the application of this tuning strategy. The analysis of the tuned preconditioner is carried out for both Hermitian and non-Hermitian eigenproblems. We show how the preconditioner can be implemented efficiently and illustrate its performance using various numerical examples. An equivalence result between the preconditioned simplified Jacobi-Davidson method and inexact inverse iteration with the tuned preconditioner is given. Finally, we discuss the shift-invert Arnoldi method both in the standard and restarted fashion. First, existing relaxation strategies for the outer iterative solves are extended to implicitly restarted Arnoldi's method. Second, we apply the idea of tuning the preconditioner to the inner iterative solve. As for inexact inverse iteration the tuned preconditioner for inexact Arnoldi's method is shown to provide significant savings in the number of inner solves. The theory in this thesis is supported by many numerical examples.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
- …