10 research outputs found

    Quasi-optimal multiplication of linear differential operators

    Get PDF
    We show that linear differential operators with polynomial coefficients over a field of characteristic zero can be multiplied in quasi-optimal time. This answers an open question raised by van der Hoeven.Comment: To appear in the Proceedings of the 53rd Annual IEEE Symposium on Foundations of Computer Science (FOCS'12

    Fast Computation of Common Left Multiples of Linear Ordinary Differential Operators

    Full text link
    We study tight bounds and fast algorithms for LCLMs of several linear differential operators with polynomial coefficients. We analyze the arithmetic complexity of existing algorithms for LCLMs, as well as the size of their outputs. We propose a new algorithm that recasts the LCLM computation in a linear algebra problem on a polynomial matrix. This algorithm yields sharp bounds on the coefficient degrees of the LCLM, improving by one order of magnitude the best bounds obtained using previous algorithms. The complexity of the new algorithm is almost optimal, in the sense that it nearly matches the arithmetic size of the output.Comment: The final version will appear in Proceedings of ISSAC 201

    Fraction-free algorithm for the computation of diagonal forms matrices over Ore domains using Gr{\"o}bner bases

    Full text link
    This paper is a sequel to "Computing diagonal form and Jacobson normal form of a matrix using Groebner bases", J. of Symb. Computation, 46 (5), 2011. We present a new fraction-free algorithm for the computation of a diagonal form of a matrix over a certain non-commutative Euclidean domain over a computable field with the help of Gr\"obner bases. This algorithm is formulated in a general constructive framework of non-commutative Ore localizations of GG-algebras (OLGAs). We split the computation of a normal form of a matrix into the diagonalization and the normalization processes. Both of them can be made fraction-free. For a matrix MM over an OLGA we provide a diagonalization algorithm to compute U,VU,V and DD with fraction-free entries such that UMV=DUMV=D holds and DD is diagonal. The fraction-free approach gives us more information on the system of linear functional equations and its solutions, than the classical setup of an operator algebra with rational functions coefficients. In particular, one can handle distributional solutions together with, say, meromorphic ones. We investigate Ore localizations of common operator algebras over K[x]K[x] and use them in the unimodularity analysis of transformation matrices U,VU,V. In turn, this allows to lift the isomorphism of modules over an OLGA Euclidean domain to a polynomial subring of it. We discuss the relation of this lifting with the solutions of the original system of equations. Moreover, we prove some new results concerning normal forms of matrices over non-simple domains. Our implementation in the computer algebra system {\sc Singular:Plural} follows the fraction-free strategy and shows impressive performance, compared with methods which directly use fractions. Since we experience moderate swell of coefficients and obtain simple transformation matrices, the method we propose is well suited for solving nontrivial practical problems.Comment: 25 pages, to appear in Journal of Symbolic Computatio

    Computing solutions of linear Mahler equations

    Get PDF
    International audienceMahler equations relate evaluations of the same function ff at iterated bbth powers of the variable. They arise in particular in the study of automatic sequences and in the complexity analysis of divide-and-conquer algorithms. Recently, the problem of solving Mahler equations in closed form has occurred in connection with number-theoretic questions. A difficulty in the manipulation of Mahler equations is the exponential blow-up of degrees when applying a Mahler operator to a polynomial. In this work, we present algorithms for solving linear Mahler equations for series, polynomials, and rational functions, and get polynomial-time complexity under a mild assumption. Incidentally, we develop an algorithm for computing the gcrd of a family of linear Mahler operators

    Matrix Polynomials and their Lower Rank Approximations

    Get PDF
    This thesis is a wide ranging work on computing a “lower-rank” approximation of a matrix polynomial using second-order non-linear optimization techniques. Two notions of rank are investigated. The first is the rank as the number of linearly independent rows or columns, which is the classical definition. The other notion considered is the lowest rank of a matrix polynomial when evaluated at a complex number, or the McCoy rank. Together, these two notions of rank allow one to compute a nearby matrix polynomial where the structure of both the left and right kernels is prescribed, along with the structure of both the infinite and finite eigenvalues. The computational theory of the calculus of matrix polynomial valued functions is developed and used in optimization algorithms based on second-order approximations. Special functions studied with a detailed error analysis are the determinant and adjoint of matrix polynomials. The unstructured and structured variants of matrix polynomials are studied in a very general setting in the context of an equality constrained optimization problem. The most general instances of these optimization problems are NP hard to approximate solutions to in a global setting. In most instances we are able to prove that solutions to our optimization problems exist (possibly at infinity) and discuss techniques in conjunction with an implementation to compute local minimizers to the problem. Most of the analysis of these problems is local and done through the Karush-Kuhn-Tucker optimality conditions for constrained optimization problems. We show that most formulations of the problems studied satisfy regularity conditions and admit Lagrange multipliers. Furthermore, we show that under some formulations that the second-order sufficient condition holds for instances of interest of the optimization problems in question. When Lagrange multipliers do not exist, we discuss why, and if it is reasonable to do so, how to regularize the problem. In several instances closed form expressions for the derivatives of matrix polynomial valued functions are derived to assist in analysis of the optimality conditions around a solution. From this analysis it is shown that variants of Newton’s method will have a local rate of convergence that is quadratic with a suitable initial guess for many problems. The implementations are demonstrated on some examples from the literature and several examples are cross-validated with different optimization formulations of the same mathematical problem. We conclude with a special application of the theory developed in this thesis is computing a nearby pair of differential polynomials with a non-trivial greatest common divisor, a non-commutative symbolic-numeric computation problem. We formulate this problem as finding a nearby structured matrix polynomial that is rank deficient in the classical sense

    Algorithmes rapides pour les polynÎmes, séries formelles et matrices

    Get PDF
    Notes d'un cours dispensĂ© aux JournĂ©es Nationales du Calcul Formel 2010International audienceLe calcul formel calcule des objets mathĂ©matiques exacts. Ce cours explore deux directions : la calculabilitĂ© et la complexitĂ©. La calculabilitĂ© Ă©tudie les classes d'objets mathĂ©matiques sur lesquelles des rĂ©ponses peuvent ĂȘtre obtenues algorithmiquement. La complexitĂ© donne ensuite des outils pour comparer des algorithmes du point de vue de leur efficacitĂ©. Ce cours passe en revue l'algorithmique efficace sur les objets fondamentaux que sont les entiers, les polynĂŽmes, les matrices, les sĂ©ries et les solutions d'Ă©quations diffĂ©rentielles ou de rĂ©currences linĂ©aires. On y montre que de nombreuses questions portant sur ces objets admettent une rĂ©ponse en complexitĂ© (quasi-)optimale, en insistant sur les principes gĂ©nĂ©raux de conception d'algorithmes efficaces. Ces notes sont dĂ©rivĂ©es du cours " Algorithmes efficaces en calcul formel " du Master Parisien de Recherche en Informatique (2004-2010), co-Ă©crit avec FrĂ©dĂ©ric Chyzak, Marc Giusti, Romain Lebreton, Bruno Salvy et Éric Schost. Le support de cours complet est disponible Ă  l'url https://wikimpri.dptinfo.ens-cachan.fr/doku.php?id=cours:c-2-2
    corecore