14 research outputs found

    Eigenvalue Methods for Interpolation Bases

    Get PDF
    This thesis investigates eigenvalue techniques for the location of roots of polynomials expressed in the Lagrange basis. Polynomial approximations to functions arise in almost all areas of computational mathematics, since polynomial expressions can be manipulated in ways that the original function cannot. Polynomials are most often expressed in the monomial basis; however, in many applications polynomials are constructed by interpolating data at a series ofpoints. The roots of such polynomial interpolants can be found by computing the eigenvalues of a generalized companion matrix pair constructed directly from the values of the interpolant. This affords the opportunity to work with polynomials expressed directly in the interpolation basis in which they were posed, avoiding the often ill-conditioned transformation between bases. Working within this framework, this thesis demonstrates that computing the roots of polynomials via these companion matrices is numerically stable, and the matrices involved can be reduced in such a way as to significantly lower the number of operations required to obtain the roots. Through examination of these various techniques, this thesis offers insight into the speed, stability, and accuracy of rootfinding algorithms for polynomials expressed in alternative bases

    Algebraic Companions and Linearizations

    Get PDF
    In this thesis, we look at a novel way of finding roots of a scalar polynomial using eigenvalue techniques. We extended this novel method to the polynomial eigenvalue problem (PEP). PEP have been used in many science and engineering applications such vibrations of structures, computer-aided geometric design, robotics, and machine learning. This thesis explains this idea in the order of which we discovered it. In Chapter 2, a new kind of companion matrix is introduced for scalar polynomials of the form c(λ)=λa(λ)b(λ)+c0c(\lambda) = \lambda a(\lambda)b(\lambda)+c_0, where upper Hessenberg companions are known for the polynomials a(λ)a(\lambda) and b(λ)b(\lambda). This construction can generate companion matrices with smaller entries than the Fiedler or Frobenius forms. This generalizes Piers Lawrence\u27s Mandelbrot companion matrix. The construction was motivated by use of Narayana-Mandelbrot polynomials. In Chapter 3, we define Euclid polynomials Ek+1(λ)=Ek(λ)(Ek(λ)1)+1E_{k+1}(\lambda) = E_{k} (\lambda) (E_{k} (\lambda) - 1) + 1 where E1(λ)=λ+1E_{1}(\lambda) = \lambda + 1 in analogy to Euclid numbers ek=Ek(1)e_k = E_{k} (1). We show how to construct companion matrices EkE_{k}, so Ek(λ)=det(λIEk)E_{k} (\lambda) = \det(\lambda I - E_{k} ) is of height 1 (and thus of minimal height over all integer companion matrices for Ek(λ)E_{k}(\lambda)). We prove various properties of these objects, and give experimental confirmation of some unproved properties. In Chapter 4, we show how to construct linearizations of matrix polynomials z\mat{a}(z)\mat{d}_0 + \mat{c}_0, \mat{a}(z)\mat{b}(z), \mat{a}(z) + \mat{b}(z) (when \deg(\mat{b}(z)) \u3c \deg(\mat{a}(z))), and z\mat{a}(z)\mat{d}_0\mat{b}(z) + \mat{c}_0 from linearizations of the component parts, matrix polynomials \mat{a}(z) and \mat{b}(z). This extends the new companion matrix construction introduced in Chapter 2 to matrix polynomials. In Chapter 5, we define ``generalized standard triples\u27\u27 which can be used in constructing algebraic linearizations; for example, for \H(z) = z \mat{a}(z)\mat{b}(z) + \mat{c}_0 from linearizations for \mat{a}(z) and \mat{b}(z). For convenience, we tabulate generalized standard triples for orthogonal polynomial bases, the monomial basis, and Newton interpolational bases; for the Bernstein basis; for Lagrange interpolational bases; and for Hermite interpolational bases. We account for the possibility of common similarity transformations. We give explicit proofs for the less familiar bases. In Chapter 6, we investigate the numerical stability of algebraic linearization, which re-uses linearizations of matrix polynomials \mat{a}(\lambda) and \mat{b}(\lambda) to make a linearization for the matrix polynomial \mat{P}(\lambda) = \lambda \mat{a}(\lambda)\mat{b}(\lambda) + \mat{c}. Such a re-use \textsl{seems} more likely to produce a well-conditioned linearization, and thus the implied algorithm for finding the eigenvalues of \mat{P}(\lambda) seems likely to be more numerically stable than expanding out the product \mat{a}(\lambda)\mat{b}(\lambda) (in whatever polynomial basis one is using). We investigate this question experimentally by using pseudospectra

    Hybrid Symbolic-Numeric Computing in Linear and Polynomial Algebra

    Get PDF
    In this thesis, we introduce hybrid symbolic-numeric methods for solving problems in linear and polynomial algebra. We mainly address the approximate GCD problem for polynomials, and problems related to parametric and polynomial matrices. For symbolic methods, our main concern is their complexity and for the numerical methods we are more concerned about their stability. The thesis consists of 5 articles which are presented in the following order: Chapter 1, deals with the fundamental notions of conditioning and backward error. Although our results are not novel, this chapter is a novel explication of conditioning and backward error that underpins the rest of the thesis. In Chapter 2, we adapt Victor Y. Pan\u27s root-based algorithm for finding approximate GCD to the case where the polynomials are expressed in Bernstein bases. We use the numerically stable companion pencil of G. F. Jónsson to compute the roots, and the Hopcroft-Karp bipartite matching method to find the degree of the approximate GCD. We offer some refinements to improve the process. In Chapter 3, we give an algorithm with similar idea to Chapter 2, which finds an approximate GCD for a pair of approximate polynomials given in a Lagrange basis. More precisely, we suppose that these polynomials are given by their approximate values at distinct known points. We first find each of their roots by using a Lagrange basis companion matrix for each polynomial. We introduce new clustering algorithms and use them to cluster the roots of each polynomial to identify multiple roots, and then marry the two polynomials using a Maximum Weight Matching (MWM) algorithm, to find their GCD. In Chapter 4, we define ``generalized standard triples\u27\u27 X, zC1 - C0, Y of regular matrix polynomials P(z) in order to use the representation X(zC1 - C0)-1 Y=P-1(z). This representation can be used in constructing algebraic linearizations; for example, for H(z) = z A(z)B(z) + C from linearizations for A(z) and B(z). This can be done even if A(z) and B(z) are expressed in differing polynomial bases. Our main theorem is that X can be expressed using the coefficients of 1 in terms of the relevant polynomial basis. For convenience we tabulate generalized standard triples for orthogonal polynomial bases, the monomial basis, and Newton interpolational bases; for the Bernstein basis; for Lagrange interpolational bases; and for Hermite interpolational bases. We account for the possibility of common similarity transformations. We give explicit proofs for the less familiar bases. Chapter 5 is devoted to parametric linear systems (PLS) and related problems, from a symbolic computational point of view. PLS are linear systems of equations in which some symbolic parameters, that is, symbols that are not considered to be candidates for elimination or solution in the course of analyzing the problem, appear in the coefficients of the system. We assume that the symbolic parameters appear polynomially in the coefficients and that the only variables to be solved for are those of the linear system. It is well-known that it is possible to specify a covering set of regimes, each of which is a semi-algebraic condition on the parameters together with a solution description valid under that condition.We provide a method of solution that requires time polynomial in the matrix dimension and the degrees of the polynomials when there are up to three parameters. Our approach exploits the Hermite and Smith normal forms that may be computed when the system coefficient domain is mapped to the univariate polynomial domain over suitably constructed fields. Our approach effectively identifies intrinsic singularities and ramification points where the algebraic and geometric structure of the matrix changes. Specially parametric eigenvalue problems can be addressed as well. Although we do not directly address the problem of computing the Jordan form, our approach allows the construction of the algebraic and geometric eigenvalue multiplicities revealed by the Frobenius form, which is a key step in the construction of the Jordan form of a matrix

    Author index for volumes 101–200

    Get PDF

    Structure-Preserving Model Reduction for Mechanical Systems

    Get PDF

    Ahlfors circle maps and total reality: from Riemann to Rohlin

    Full text link
    This is a prejudiced survey on the Ahlfors (extremal) function and the weaker {\it circle maps} (Garabedian-Schiffer's translation of "Kreisabbildung"), i.e. those (branched) maps effecting the conformal representation upon the disc of a {\it compact bordered Riemann surface}. The theory in question has some well-known intersection with real algebraic geometry, especially Klein's ortho-symmetric curves via the paradigm of {\it total reality}. This leads to a gallery of pictures quite pleasant to visit of which we have attempted to trace the simplest representatives. This drifted us toward some electrodynamic motions along real circuits of dividing curves perhaps reminiscent of Kepler's planetary motions along ellipses. The ultimate origin of circle maps is of course to be traced back to Riemann's Thesis 1851 as well as his 1857 Nachlass. Apart from an abrupt claim by Teichm\"uller 1941 that everything is to be found in Klein (what we failed to assess on printed evidence), the pivotal contribution belongs to Ahlfors 1950 supplying an existence-proof of circle maps, as well as an analysis of an allied function-theoretic extremal problem. Works by Yamada 1978--2001, Gouma 1998 and Coppens 2011 suggest sharper degree controls than available in Ahlfors' era. Accordingly, our partisan belief is that much remains to be clarified regarding the foundation and optimal control of Ahlfors circle maps. The game of sharp estimation may look narrow-minded "Absch\"atzungsmathematik" alike, yet the philosophical outcome is as usual to contemplate how conformal and algebraic geometry are fighting together for the soul of Riemann surfaces. A second part explores the connection with Hilbert's 16th as envisioned by Rohlin 1978.Comment: 675 pages, 199 figures; extended version of the former text (v.1) by including now Rohlin's theory (v.2
    corecore