10,093 research outputs found

    Fast Computation of Common Left Multiples of Linear Ordinary Differential Operators

    Full text link
    We study tight bounds and fast algorithms for LCLMs of several linear differential operators with polynomial coefficients. We analyze the arithmetic complexity of existing algorithms for LCLMs, as well as the size of their outputs. We propose a new algorithm that recasts the LCLM computation in a linear algebra problem on a polynomial matrix. This algorithm yields sharp bounds on the coefficient degrees of the LCLM, improving by one order of magnitude the best bounds obtained using previous algorithms. The complexity of the new algorithm is almost optimal, in the sense that it nearly matches the arithmetic size of the output.Comment: The final version will appear in Proceedings of ISSAC 201

    Fine-grained dichotomies for the Tutte plane and Boolean #CSP

    Get PDF
    Jaeger, Vertigan, and Welsh [15] proved a dichotomy for the complexity of evaluating the Tutte polynomial at fixed points: The evaluation is #P-hard almost everywhere, and the remaining points admit polynomial-time algorithms. Dell, Husfeldt, and Wahl\'en [9] and Husfeldt and Taslaman [12], in combination with Curticapean [7], extended the #P-hardness results to tight lower bounds under the counting exponential time hypothesis #ETH, with the exception of the line y=1y=1, which was left open. We complete the dichotomy theorem for the Tutte polynomial under #ETH by proving that the number of all acyclic subgraphs of a given nn-vertex graph cannot be determined in time exp(o(n))exp(o(n)) unless #ETH fails. Another dichotomy theorem we strengthen is the one of Creignou and Hermann [6] for counting the number of satisfying assignments to a constraint satisfaction problem instance over the Boolean domain. We prove that all #P-hard cases are also hard under #ETH. The main ingredient is to prove that the number of independent sets in bipartite graphs with nn vertices cannot be computed in time exp(o(n))exp(o(n)) unless #ETH fails. In order to prove our results, we use the block interpolation idea by Curticapean [7] and transfer it to systems of linear equations that might not directly correspond to interpolation.Comment: 16 pages, 1 figur

    Polynomial interpolation and Gaussian quadrature for matrix valued functions

    Get PDF
    The techniques for polynomial interpolation and Gaussian quadrature are generalized to matrix-valued functions. It is shown how the zeros and rootvectors of matrix orthonormal polynomials can be used to get a quadrature formula with the highest degree of precision
    • …
    corecore