1,922 research outputs found

    Forest matrices around the Laplacian matrix

    Get PDF
    We study the matrices Q_k of in-forests of a weighted digraph G and their connections with the Laplacian matrix L of G. The (i,j) entry of Q_k is the total weight of spanning converging forests (in-forests) with k arcs such that i belongs to a tree rooted at j. The forest matrices, Q_k, can be calculated recursively and expressed by polynomials in the Laplacian matrix; they provide representations for the generalized inverses, the powers, and some eigenvectors of L. The normalized in-forest matrices are row stochastic; the normalized matrix of maximum in-forests is the eigenprojection of the Laplacian matrix, which provides an immediate proof of the Markov chain tree theorem. A source of these results is the fact that matrices Q_k are the matrix coefficients in the polynomial expansion of adj(a*I+L). Thereby they are precisely Faddeev's matrices for -L. Keywords: Weighted digraph; Laplacian matrix; Spanning forest; Matrix-forest theorem; Leverrier-Faddeev method; Markov chain tree theorem; Eigenprojection; Generalized inverse; Singular M-matrixComment: 19 pages, presented at the Edinburgh (2001) Conference on Algebraic Graph Theor

    The asymptotics a Bessel-kernel determinant which arises in Random Matrix Theory

    Full text link
    In Random Matrix Theory the local correlations of the Laguerre and Jacobi Unitary Ensemble in the hard edge scaling limit can be described in terms of the Bessel kernel (containing a parameter α\alpha). In particular, the so-called hard edge gap probabilities can be expressed as the Fredholm determinants of the corresponding integral operator restricted to the finite interval [0, R]. Using operator theoretic methods we are going to compute their asymptotics as R goes to infinity under certain assumption on the parameter α\alpha.Comment: 50 page

    Asymptotics of block Toeplitz determinants and the classical dimer model

    Full text link
    We compute the asymptotics of a block Toeplitz determinant which arises in the classical dimer model for the triangular lattice when considering the monomer-monomer correlation function. The model depends on a parameter interpolating between the square lattice (t=0t=0) and the triangular lattice (t=1t=1), and we obtain the asymptotics for 0<t≤10<t\le 1. For 0<t<10<t<1 we apply the Szeg\"o Limit Theorem for block Toeplitz determinants. The main difficulty is to evaluate the constant term in the asymptotics, which is generally given only in a rather abstract form

    Numeric and symbolic evaluation of the pfaffian of general skew-symmetric matrices

    Full text link
    Evaluation of pfaffians arises in a number of physics applications, and for some of them a direct method is preferable to using the determinantal formula. We discuss two methods for the numerical evaluation of pfaffians. The first is tridiagonalization based on Householder transformations. The main advantage of this method is its numerical stability that makes unnecessary the implementation of a pivoting strategy. The second method considered is based on Aitken's block diagonalization formula. It yields to a kind of LU (similar to Cholesky's factorization) decomposition (under congruence) of arbitrary skew-symmetric matrices that is well suited both for the numeric and symbolic evaluations of the pfaffian. Fortran subroutines (FORTRAN 77 and 90) implementing both methods are given. We also provide simple implementations in Python and Mathematica for purpose of testing, or for exploratory studies of methods that make use of pfaffians.Comment: 13 pages, Download links: http://gamma.ft.uam.es/robledo/Downloads.html and http://www.phys.washington.edu/users/bertsch/computer.htm

    Accurate and Efficient Expression Evaluation and Linear Algebra

    Full text link
    We survey and unify recent results on the existence of accurate algorithms for evaluating multivariate polynomials, and more generally for accurate numerical linear algebra with structured matrices. By "accurate" we mean that the computed answer has relative error less than 1, i.e., has some correct leading digits. We also address efficiency, by which we mean algorithms that run in polynomial time in the size of the input. Our results will depend strongly on the model of arithmetic: Most of our results will use the so-called Traditional Model (TM). We give a set of necessary and sufficient conditions to decide whether a high accuracy algorithm exists in the TM, and describe progress toward a decision procedure that will take any problem and provide either a high accuracy algorithm or a proof that none exists. When no accurate algorithm exists in the TM, it is natural to extend the set of available accurate operations by a library of additional operations, such as x+y+zx+y+z, dot products, or indeed any enumerable set which could then be used to build further accurate algorithms. We show how our accurate algorithms and decision procedure for finding them extend to this case. Finally, we address other models of arithmetic, and the relationship between (im)possibility in the TM and (in)efficient algorithms operating on numbers represented as bit strings.Comment: 49 pages, 6 figures, 1 tabl
    • …
    corecore