161 research outputs found
Pseudorandomness for Approximate Counting and Sampling
We study computational procedures that use both randomness and nondeterminism. The goal of this paper is to derandomize such procedures under the weakest possible assumptions.
Our main technical contribution allows one to “boost” a given hardness assumption: We show that if there is a problem in EXP that cannot be computed by poly-size nondeterministic circuits then there is one which cannot be computed by poly-size circuits that make non-adaptive NP oracle queries. This in particular shows that the various assumptions used over the last few years by several authors to derandomize Arthur-Merlin games (i.e., show AM = NP) are in fact all equivalent.
We also define two new primitives that we regard as the natural pseudorandom objects associated with approximate counting and sampling of NP-witnesses. We use the “boosting” theorem and hashing techniques to construct these primitives using an assumption that is no stronger than that used to derandomize AM.
We observe that Cai's proof that S_2^P ⊆ PP⊆(NP) and the learning algorithm of Bshouty et al. can be seen as reductions to sampling that are not probabilistic. As a consequence they can be derandomized under an assumption which is weaker than the assumption that was previously known to suffice
Complete Problems for Multi-Pseudodeterministic Computations
We exhibit several computational problems that are complete for multi-pseudodeterministic computations in the following sense: (1) these problems admit 2-pseudodeterministic algorithms (2) if there exists a pseudodeterministic algorithm for any of these problems, then any multi-valued function that admits a k-pseudodeterministic algorithm for a constant k, also admits a pseudodeterministic algorithm. We also show that these computational problems are complete for Search-BPP: a pseudodeterministic algorithm for any of these problems implies a pseudodeterministic algorithm for all problems in Search-BPP
Derandomizing Compressed Sensing With Combinatorial Design
Compressed sensing is the art of effectively reconstructing structured n-dimensional vectors from substantially fewer measurements than naively anticipated. A plethora of analytical reconstruction guarantees support this credo. The strongest among them are based on deep results from large-dimensional probability theory and require a considerable amount of randomness in the measurement design. Here, we demonstrate that derandomization techniques allow for a considerable reduction in the randomness required for such proof strategies. More precisely, we establish uniform s-sparse reconstruction guarantees for Cs log(n) measurements that are chosen independently from strength-4 orthogonal arrays and maximal sets of mutually unbiased bases, respectively. These are highly structured families of Ĉn2 vectors that imitate signed Bernoulli and standard Gaussian vectors in a (partially) derandomized fashion.DFG, 414044773, Open Access Publizieren 2019 - 2020 / Technische Universität Berli
Jacobian hits circuits: Hitting-sets, lower bounds for depth-D occur-k formulas & depth-3 transcendence degree-k circuits
We present a single, common tool to strictly subsume all known cases of
polynomial time blackbox polynomial identity testing (PIT) that have been
hitherto solved using diverse tools and techniques. In particular, we show that
polynomial time hitting-set generators for identity testing of the two
seemingly different and well studied models - depth-3 circuits with bounded top
fanin, and constant-depth constant-read multilinear formulas - can be
constructed using one common algebraic-geometry theme: Jacobian captures
algebraic independence. By exploiting the Jacobian, we design the first
efficient hitting-set generators for broad generalizations of the
above-mentioned models, namely:
(1) depth-3 (Sigma-Pi-Sigma) circuits with constant transcendence degree of
the polynomials computed by the product gates (no bounded top fanin
restriction), and (2) constant-depth constant-occur formulas (no multilinear
restriction).
Constant-occur of a variable, as we define it, is a much more general concept
than constant-read. Also, earlier work on the latter model assumed that the
formula is multilinear. Thus, our work goes further beyond the results obtained
by Saxena & Seshadhri (STOC 2011), Saraf & Volkovich (STOC 2011), Anderson et
al. (CCC 2011), Beecken et al. (ICALP 2011) and Grenet et al. (FSTTCS 2011),
and brings them under one unifying technique.
In addition, using the same Jacobian based approach, we prove exponential
lower bounds for the immanant (which includes permanent and determinant) on the
same depth-3 and depth-4 models for which we give efficient PIT algorithms. Our
results reinforce the intimate connection between identity testing and lower
bounds by exhibiting a concrete mathematical tool - the Jacobian - that is
equally effective in solving both the problems on certain interesting and
previously well-investigated (but not well understood) models of computation
Pseudo-Deterministic Streaming
A pseudo-deterministic algorithm is a (randomized) algorithm which, when run multiple times on the same input, with high probability outputs the same result on all executions. Classic streaming algorithms, such as those for finding heavy hitters, approximate counting, ?_2 approximation, finding a nonzero entry in a vector (for turnstile algorithms) are not pseudo-deterministic. For example, in the instance of finding a nonzero entry in a vector, for any known low-space algorithm A, there exists a stream x so that running A twice on x (using different randomness) would with high probability result in two different entries as the output.
In this work, we study whether it is inherent that these algorithms output different values on different executions. That is, we ask whether these problems have low-memory pseudo-deterministic algorithms. For instance, we show that there is no low-memory pseudo-deterministic algorithm for finding a nonzero entry in a vector (given in a turnstile fashion), and also that there is no low-dimensional pseudo-deterministic sketching algorithm for ?_2 norm estimation. We also exhibit problems which do have low memory pseudo-deterministic algorithms but no low memory deterministic algorithm, such as outputting a nonzero row of a matrix, or outputting a basis for the row-span of a matrix.
We also investigate multi-pseudo-deterministic algorithms: algorithms which with high probability output one of a few options. We show the first lower bounds for such algorithms. This implies that there are streaming problems such that every low space algorithm for the problem must have inputs where there are many valid outputs, all with a significant probability of being outputted
Deterministically Counting Satisfying Assignments for Constant-Depth Circuits with Parity Gates, with Implications for Lower Bounds
We give a deterministic algorithm for counting the number of satisfying assignments of any AC^0[oplus] circuit C of size s and depth d over n variables in time 2^(n-f(n,s,d)), where f(n,s,d) = n/O(log(s))^(d-1), whenever s = 2^o(n^(1/d)). As a consequence, we get that for each d, there is a language in E^{NP} that does not have AC^0[oplus] circuits of size 2^o(n^(1/(d+1))). This is the first lower bound in E^{NP} against AC^0[oplus] circuits that beats the lower bound of 2^Omega(n^(1/2(d-1))) due to Razborov and Smolensky for large d. Both our algorithm and our lower bounds extend to AC^0[p] circuits for any prime p
- …