21 research outputs found

    Complete Problems for Multi-Pseudodeterministic Computations

    Get PDF
    We exhibit several computational problems that are complete for multi-pseudodeterministic computations in the following sense: (1) these problems admit 2-pseudodeterministic algorithms (2) if there exists a pseudodeterministic algorithm for any of these problems, then any multi-valued function that admits a k-pseudodeterministic algorithm for a constant k, also admits a pseudodeterministic algorithm. We also show that these computational problems are complete for Search-BPP: a pseudodeterministic algorithm for any of these problems implies a pseudodeterministic algorithm for all problems in Search-BPP

    Pseudodeterministic constructions in subexponential time

    Get PDF
    We study pseudodeterministic constructions, i.e., randomized algorithms which output the same solution on most computation paths. We establish unconditionally that there is an infinite sequence {pn}n∈N of increasing primes and a randomized algorithm A running in expected sub-exponential time such that for each n, on input 1|pn|, A outputs pn with probability 1. In other words, our result provides a pseudodeterministic construction of primes in sub-exponential time which works infinitely often. This result follows from a much more general theorem about pseudodeterministic constructions. A property Q ⊆ {0, 1}* is γ-dense if for large enough n, |Q ⋂ {0, 1}n| ≥ γ2n. We show that for each c > 0 at least one of the following holds: (1) There is a pseudodeterministic polynomial time construction of a family {Hn} of sets, Hn ⊆ {0, 1}n, such that for each (1=nc)-dense property Q ∈ DTIME(n^c) and every large enough n, Hn ⋂ Q ≠ ∅; or (2) There is a deterministic sub-exponential time construction of a family {H'n} of sets, H'n ⊆ {0, 1}n, such that for each (1/n^c)-dense property Q ∈ DTIME(n^c) and for infinitely many values of n, H'n ⋂ Q ≠ ∅. We provide further algorithmic applications that might be of independent interest. Perhaps intriguingly, while our main results are unconditional, they have a non-constructive element, arising from a sequence of applications of the hardness versus randomness paradigm.</p

    Geometry of Rounding: Near Optimal Bounds and a New Neighborhood Sperner's Lemma

    Full text link
    A partition P\mathcal{P} of Rd\mathbb{R}^d is called a (k,ε)(k,\varepsilon)-secluded partition if, for every pRd\vec{p} \in \mathbb{R}^d, the ball B(ε,p)\overline{B}_{\infty}(\varepsilon, \vec{p}) intersects at most kk members of P\mathcal{P}. A goal in designing such secluded partitions is to minimize kk while making ε\varepsilon as large as possible. This partition problem has connections to a diverse range of topics, including deterministic rounding schemes, pseudodeterminism, replicability, as well as Sperner/KKM-type results. In this work, we establish near-optimal relationships between kk and ε\varepsilon. We show that, for any bounded measure partitions and for any d1d\geq 1, it must be that k(1+2ε)dk\geq(1+2\varepsilon)^d. Thus, when k=k(d)k=k(d) is restricted to poly(d){\rm poly}(d), it follows that ε=ε(d)O(lndd)\varepsilon=\varepsilon(d)\in O\left(\frac{\ln d}{d}\right). This bound is tight up to log factors, as it is known that there exist secluded partitions with k(d)=d+1k(d)=d+1 and ε(d)=12d\varepsilon(d)=\frac{1}{2d}. We also provide new constructions of secluded partitions that work for a broad spectrum of k(d)k(d) and ε(d)\varepsilon(d) parameters. Specifically, we prove that, for any f:NNf:\mathbb{N}\rightarrow\mathbb{N}, there is a secluded partition with k(d)=(f(d)+1)df(d)k(d)=(f(d)+1)^{\lceil\frac{d}{f(d)}\rceil} and ε(d)=12f(d)\varepsilon(d)=\frac{1}{2f(d)}. These new partitions are optimal up to O(logd)O(\log d) factors for various choices of k(d)k(d) and ε(d)\varepsilon(d). Based on the lower bound result, we establish a new neighborhood version of Sperner's lemma over hypercubes, which is of independent interest. In addition, we prove a no-free-lunch theorem about the limitations of rounding schemes in the context of pseudodeterministic/replicable algorithms

    Polynomial-Time Pseudodeterministic Construction of Primes

    Full text link
    A randomized algorithm for a search problem is *pseudodeterministic* if it produces a fixed canonical solution to the search problem with high probability. In their seminal work on the topic, Gat and Goldwasser posed as their main open problem whether prime numbers can be pseudodeterministically constructed in polynomial time. We provide a positive solution to this question in the infinitely-often regime. In more detail, we give an *unconditional* polynomial-time randomized algorithm BB such that, for infinitely many values of nn, B(1n)B(1^n) outputs a canonical nn-bit prime pnp_n with high probability. More generally, we prove that for every dense property QQ of strings that can be decided in polynomial time, there is an infinitely-often pseudodeterministic polynomial-time construction of strings satisfying QQ. This improves upon a subexponential-time construction of Oliveira and Santhanam. Our construction uses several new ideas, including a novel bootstrapping technique for pseudodeterministic constructions, and a quantitative optimization of the uniform hardness-randomness framework of Chen and Tell, using a variant of the Shaltiel--Umans generator

    On the Pseudo-Deterministic Query Complexity of NP Search Problems

    Get PDF
    We study pseudo-deterministic query complexity - randomized query algorithms that are required to output the same answer with high probability on all inputs. We prove Ω(√n) lower bounds on the pseudo-deterministic complexity of a large family of search problems based on unsatisfiable random CNF instances, and also for the promise problem (FIND1) of finding a 1 in a vector populated with at least half one’s. This gives an exponential separation between randomized query complexity and pseudo-deterministic complexity, which is tight in the quantum setting. As applications we partially solve a related combinatorial coloring problem, and we separate random tree-like Resolution from its pseudo-deterministic version. In contrast to our lower bound, we show, surprisingly, that in the zero-error, average case setting, the three notions (deterministic, randomized, pseudo-deterministic) collapse

    Geometry of Rounding

    Full text link
    Rounding has proven to be a fundamental tool in theoretical computer science. By observing that rounding and partitioning of Rd\mathbb{R}^d are equivalent, we introduce the following natural partition problem which we call the {\em secluded hypercube partition problem}: Given kNk\in \mathbb{N} (ideally small) and ϵ>0\epsilon>0 (ideally large), is there a partition of Rd\mathbb{R}^d with unit hypercubes such that for every point pRdp \in \mathbb{R}^d, its closed ϵ\epsilon-neighborhood (in the \ell_{\infty} norm) intersects at most kk hypercubes? We undertake a comprehensive study of this partition problem. We prove that for every dNd\in \mathbb{N}, there is an explicit (and efficiently computable) hypercube partition of Rd\mathbb{R}^d with k=d+1k = d+1 and ϵ=12d\epsilon = \frac{1}{2d}. We complement this construction by proving that the value of k=d+1k=d+1 is the best possible (for any ϵ\epsilon) for a broad class of ``reasonable'' partitions including hypercube partitions. We also investigate the optimality of the parameter ϵ\epsilon and prove that any partition in this broad class that has k=d+1k=d+1, must have ϵ12d\epsilon\leq\frac{1}{2\sqrt{d}}. These bounds imply limitations of certain deterministic rounding schemes existing in the literature. Furthermore, this general bound is based on the currently known lower bounds for the dissection number of the cube, and improvements to this bound will yield improvements to our bounds. While our work is motivated by the desire to understand rounding algorithms, one of our main conceptual contributions is the introduction of the {\em secluded hypercube partition problem}, which fits well with a long history of investigations by mathematicians on various hypercube partitions/tilings of Euclidean space

    Complexity Theory

    Get PDF
    Computational Complexity Theory is the mathematical study of the intrinsic power and limitations of computational resources like time, space, or randomness. The current workshop focused on recent developments in various sub-areas including arithmetic complexity, Boolean complexity, communication complexity, cryptography, probabilistic proof systems, pseudorandomness, and quantum computation. Many of the developments are related to diverse mathematical fields such as algebraic geometry, combinatorial number theory, probability theory, representation theory, and the theory of error-correcting codes

    Planar Maximum Matching: Towards a Parallel Algorithm

    Get PDF
    corecore