4,512 research outputs found

    Efficiently Computing Real Roots of Sparse Polynomials

    Full text link
    We propose an efficient algorithm to compute the real roots of a sparse polynomial fR[x]f\in\mathbb{R}[x] having kk non-zero real-valued coefficients. It is assumed that arbitrarily good approximations of the non-zero coefficients are given by means of a coefficient oracle. For a given positive integer LL, our algorithm returns disjoint disks Δ1,,ΔsC\Delta_{1},\ldots,\Delta_{s}\subset\mathbb{C}, with s<2ks<2k, centered at the real axis and of radius less than 2L2^{-L} together with positive integers μ1,,μs\mu_{1},\ldots,\mu_{s} such that each disk Δi\Delta_{i} contains exactly μi\mu_{i} roots of ff counted with multiplicity. In addition, it is ensured that each real root of ff is contained in one of the disks. If ff has only simple real roots, our algorithm can also be used to isolate all real roots. The bit complexity of our algorithm is polynomial in kk and logn\log n, and near-linear in LL and τ\tau, where 2τ2^{-\tau} and 2τ2^{\tau} constitute lower and upper bounds on the absolute values of the non-zero coefficients of ff, and nn is the degree of ff. For root isolation, the bit complexity is polynomial in kk and logn\log n, and near-linear in τ\tau and logσ1\log\sigma^{-1}, where σ\sigma denotes the separation of the real roots

    Shallow Circuits with High-Powered Inputs

    Get PDF
    A polynomial identity testing algorithm must determine whether an input polynomial (given for instance by an arithmetic circuit) is identically equal to 0. In this paper, we show that a deterministic black-box identity testing algorithm for (high-degree) univariate polynomials would imply a lower bound on the arithmetic complexity of the permanent. The lower bounds that are known to follow from derandomization of (low-degree) multivariate identity testing are weaker. To obtain our lower bound it would be sufficient to derandomize identity testing for polynomials of a very specific norm: sums of products of sparse polynomials with sparse coefficients. This observation leads to new versions of the Shub-Smale tau-conjecture on integer roots of univariate polynomials. In particular, we show that a lower bound for the permanent would follow if one could give a good enough bound on the number of real roots of sums of products of sparse polynomials (Descartes' rule of signs gives such a bound for sparse polynomials and products thereof). In this third version of our paper we show that the same lower bound would follow even if one could only prove a slightly superpolynomial upper bound on the number of real roots. This is a consequence of a new result on reduction to depth 4 for arithmetic circuits which we establish in a companion paper. We also show that an even weaker bound on the number of real roots would suffice to obtain a lower bound on the size of depth 4 circuits computing the permanent.Comment: A few typos correcte

    Efficiently Detecting Torsion Points and Subtori

    Full text link
    Suppose X is the complex zero set of a finite collection of polynomials in Z[x_1,...,x_n]. We show that deciding whether X contains a point all of whose coordinates are d_th roots of unity can be done within NP^NP (relative to the sparse encoding), under a plausible assumption on primes in arithmetic progression. In particular, our hypothesis can still hold even under certain failures of the Generalized Riemann Hypothesis, such as the presence of Siegel-Landau zeroes. Furthermore, we give a similar (but UNconditional) complexity upper bound for n=1. Finally, letting T be any algebraic subgroup of (C^*)^n we show that deciding whether X contains T is coNP-complete (relative to an even more efficient encoding),unconditionally. We thus obtain new non-trivial families of multivariate polynomial systems where deciding the existence of complex roots can be done unconditionally in the polynomial hierarchy -- a family of complexity classes lying between PSPACE and P, intimately connected with the P=?NP Problem. We also discuss a connection to Laurent's solution of Chabauty's Conjecture from arithmetic geometry.Comment: 21 pages, no figures. Final version, with additional commentary and references. Also fixes a gap in Theorems 2 (now Theorem 1.3) regarding translated subtor

    A Near-Optimal Algorithm for Computing Real Roots of Sparse Polynomials

    Full text link
    Let pZ[x]p\in\mathbb{Z}[x] be an arbitrary polynomial of degree nn with kk non-zero integer coefficients of absolute value less than 2τ2^\tau. In this paper, we answer the open question whether the real roots of pp can be computed with a number of arithmetic operations over the rational numbers that is polynomial in the input size of the sparse representation of pp. More precisely, we give a deterministic, complete, and certified algorithm that determines isolating intervals for all real roots of pp with O(k3log(nτ)logn)O(k^3\cdot\log(n\tau)\cdot \log n) many exact arithmetic operations over the rational numbers. When using approximate but certified arithmetic, the bit complexity of our algorithm is bounded by O~(k4nτ)\tilde{O}(k^4\cdot n\tau), where O~()\tilde{O}(\cdot) means that we ignore logarithmic. Hence, for sufficiently sparse polynomials (i.e. k=O(logc(nτ))k=O(\log^c (n\tau)) for a positive constant cc), the bit complexity is O~(nτ)\tilde{O}(n\tau). We also prove that the latter bound is optimal up to logarithmic factors

    An Elimination Method for Solving Bivariate Polynomial Systems: Eliminating the Usual Drawbacks

    Full text link
    We present an exact and complete algorithm to isolate the real solutions of a zero-dimensional bivariate polynomial system. The proposed algorithm constitutes an elimination method which improves upon existing approaches in a number of points. First, the amount of purely symbolic operations is significantly reduced, that is, only resultant computation and square-free factorization is still needed. Second, our algorithm neither assumes generic position of the input system nor demands for any change of the coordinate system. The latter is due to a novel inclusion predicate to certify that a certain region is isolating for a solution. Our implementation exploits graphics hardware to expedite the resultant computation. Furthermore, we integrate a number of filtering techniques to improve the overall performance. Efficiency of the proposed method is proven by a comparison of our implementation with two state-of-the-art implementations, that is, LPG and Maple's isolate. For a series of challenging benchmark instances, experiments show that our implementation outperforms both contestants.Comment: 16 pages with appendix, 1 figure, submitted to ALENEX 201

    Syndrome decoding of Reed-Muller codes and tensor decomposition over finite fields

    Full text link
    Reed-Muller codes are some of the oldest and most widely studied error-correcting codes, of interest for both their algebraic structure as well as their many algorithmic properties. A recent beautiful result of Saptharishi, Shpilka and Volk showed that for binary Reed-Muller codes of length nn and distance d=O(1)d = O(1), one can correct polylog(n)\operatorname{polylog}(n) random errors in poly(n)\operatorname{poly}(n) time (which is well beyond the worst-case error tolerance of O(1)O(1)). In this paper, we consider the problem of `syndrome decoding' Reed-Muller codes from random errors. More specifically, given the polylog(n)\operatorname{polylog}(n)-bit long syndrome vector of a codeword corrupted in polylog(n)\operatorname{polylog}(n) random coordinates, we would like to compute the locations of the codeword corruptions. This problem turns out to be equivalent to a basic question about computing tensor decomposition of random low-rank tensors over finite fields. Our main result is that syndrome decoding of Reed-Muller codes (and the equivalent tensor decomposition problem) can be solved efficiently, i.e., in polylog(n)\operatorname{polylog}(n) time. We give two algorithms for this problem: 1. The first algorithm is a finite field variant of a classical algorithm for tensor decomposition over real numbers due to Jennrich. This also gives an alternate proof for the main result of Saptharishi et al. 2. The second algorithm is obtained by implementing the steps of the Berlekamp-Welch-style decoding algorithm of Saptharishi et al. in sublinear-time. The main new ingredient is an algorithm for solving certain kinds of systems of polynomial equations.Comment: 24 page
    corecore