76,681 research outputs found

    The computational complexity of algebraic numbers

    Full text link

    On the computational complexity of algebraic numbers : the Hartmanis-Stearns problem revisited

    No full text
    International audience— We consider the complexity of integer base expansions of algebraic irrational numbers from a computational point of view. We show that the Hartmanis–Stearns problem can be solved in a satisfactory way for the class of multistack machines. In this direction, our main result is that the base-b expansion of an algebraic irrational real number cannot be generated by a deterministic pushdown automaton. We also confirm an old claim of Cobham proving that such numbers cannot be generated by a tag machine with dilation factor larger than one

    Query-Efficient Locally Decodable Codes of Subexponential Length

    Full text link
    We develop the algebraic theory behind the constructions of Yekhanin (2008) and Efremenko (2009), in an attempt to understand the ``algebraic niceness'' phenomenon in Zm\mathbb{Z}_m. We show that every integer m=pq=2t−1m = pq = 2^t -1, where pp, qq and tt are prime, possesses the same good algebraic property as m=511m=511 that allows savings in query complexity. We identify 50 numbers of this form by computer search, which together with 511, are then applied to gain improvements on query complexity via Itoh and Suzuki's composition method. More precisely, we construct a 3⌈r/2⌉3^{\lceil r/2\rceil}-query LDC for every positive integer r<104r<104 and a ⌊(3/4)51⋅2r⌋\left\lfloor (3/4)^{51}\cdot 2^{r}\right\rfloor-query LDC for every integer r≥104r\geq 104, both of length NrN_{r}, improving the 2r2^r queries used by Efremenko (2009) and 3⋅2r−23\cdot 2^{r-2} queries used by Itoh and Suzuki (2010). We also obtain new efficient private information retrieval (PIR) schemes from the new query-efficient LDCs.Comment: to appear in Computational Complexit

    A complex analogue of Toda's Theorem

    Full text link
    Toda \cite{Toda} proved in 1989 that the (discrete) polynomial time hierarchy, PH\mathbf{PH}, is contained in the class \mathbf{P}^{#\mathbf{P}}, namely the class of languages that can be decided by a Turing machine in polynomial time given access to an oracle with the power to compute a function in the counting complexity class #\mathbf{P}. This result, which illustrates the power of counting is considered to be a seminal result in computational complexity theory. An analogous result (with a compactness hypothesis) in the complexity theory over the reals (in the sense of Blum-Shub-Smale real machines \cite{BSS89}) was proved in \cite{BZ09}. Unlike Toda's proof in the discrete case, which relied on sophisticated combinatorial arguments, the proof in \cite{BZ09} is topological in nature in which the properties of the topological join is used in a fundamental way. However, the constructions used in \cite{BZ09} were semi-algebraic -- they used real inequalities in an essential way and as such do not extend to the complex case. In this paper, we extend the techniques developed in \cite{BZ09} to the complex projective case. A key role is played by the complex join of quasi-projective complex varieties. As a consequence we obtain a complex analogue of Toda's theorem. The results contained in this paper, taken together with those contained in \cite{BZ09}, illustrate the central role of the Poincar\'e polynomial in algorithmic algebraic geometry, as well as, in computational complexity theory over the complex and real numbers -- namely, the ability to compute it efficiently enables one to decide in polynomial time all languages in the (compact) polynomial hierarchy over the appropriate field.Comment: 31 pages. Final version to appear in Foundations of Computational Mathematic

    Synthesizing Switching Controllers for Hybrid Systems by Continuous Invariant Generation

    Full text link
    We extend a template-based approach for synthesizing switching controllers for semi-algebraic hybrid systems, in which all expressions are polynomials. This is achieved by combining a QE (quantifier elimination)-based method for generating continuous invariants with a qualitative approach for predefining templates. Our synthesis method is relatively complete with regard to a given family of predefined templates. Using qualitative analysis, we discuss heuristics to reduce the numbers of parameters appearing in the templates. To avoid too much human interaction in choosing templates as well as the high computational complexity caused by QE, we further investigate applications of the SOS (sum-of-squares) relaxation approach and the template polyhedra approach in continuous invariant generation, which are both well supported by efficient numerical solvers

    An odyssey into local refinement and multilevel preconditioning III: Implementation and numerical experiments

    Get PDF
    In this paper, we examine a number of additive and multiplicative multilevel iterative methods and preconditioners in the setting of two-dimensional local mesh refinement. While standard multilevel methods are effective for uniform refinement-based discretizations of elliptic equations, they tend to be less effective for algebraic systems, which arise from discretizations on locally refined meshes, losing their optimal behavior in both storage and computational complexity. Our primary focus here is on Bramble, Pasciak, and Xu (BPX)-style additive and multiplicative multilevel preconditioners, and on various stabilizations of the additive and multiplicative hierarchical basis (HB) method, and their use in the local mesh refinement setting. In parts I and II of this trilogy, it was shown that both BPX and wavelet stabilizations of HB have uniformly bounded condition numbers on several classes of locally refined two- and three-dimensional meshes based on fairly standard (and easily implementable) red and red-green mesh refinement algorithms. In this third part of the trilogy, we describe in detail the implementation of these types of algorithms, including detailed discussions of the data structures and traversal algorithms we employ for obtaining optimal storage and computational complexity in our implementations. We show how each of the algorithms can be implemented using standard data types, available in languages such as C and FORTRAN, so that the resulting algorithms have optimal (linear) storage requirements, and so that the resulting multilevel method or preconditioner can be applied with optimal (linear) computational costs. We have successfully used these data structure ideas for both MATLAB and C implementations using the FEtk, an open source finite element software package. We finish the paper with a sequence of numerical experiments illustrating the effectiveness of a number of BPX and stabilized HB variants for several examples requiring local refinement
    • …
    corecore