7,760 research outputs found
Shallow Circuits with High-Powered Inputs
A polynomial identity testing algorithm must determine whether an input
polynomial (given for instance by an arithmetic circuit) is identically equal
to 0. In this paper, we show that a deterministic black-box identity testing
algorithm for (high-degree) univariate polynomials would imply a lower bound on
the arithmetic complexity of the permanent. The lower bounds that are known to
follow from derandomization of (low-degree) multivariate identity testing are
weaker. To obtain our lower bound it would be sufficient to derandomize
identity testing for polynomials of a very specific norm: sums of products of
sparse polynomials with sparse coefficients. This observation leads to new
versions of the Shub-Smale tau-conjecture on integer roots of univariate
polynomials. In particular, we show that a lower bound for the permanent would
follow if one could give a good enough bound on the number of real roots of
sums of products of sparse polynomials (Descartes' rule of signs gives such a
bound for sparse polynomials and products thereof). In this third version of
our paper we show that the same lower bound would follow even if one could only
prove a slightly superpolynomial upper bound on the number of real roots. This
is a consequence of a new result on reduction to depth 4 for arithmetic
circuits which we establish in a companion paper. We also show that an even
weaker bound on the number of real roots would suffice to obtain a lower bound
on the size of depth 4 circuits computing the permanent.Comment: A few typos correcte
Consistency of circuit lower bounds with bounded theories
Proving that there are problems in that require
boolean circuits of super-linear size is a major frontier in complexity theory.
While such lower bounds are known for larger complexity classes, existing
results only show that the corresponding problems are hard on infinitely many
input lengths. For instance, proving almost-everywhere circuit lower bounds is
open even for problems in . Giving the notorious difficulty of
proving lower bounds that hold for all large input lengths, we ask the
following question: Can we show that a large set of techniques cannot prove
that is easy infinitely often? Motivated by this and related
questions about the interaction between mathematical proofs and computations,
we investigate circuit complexity from the perspective of logic.
Among other results, we prove that for any parameter it is
consistent with theory that computational class , where is one of
the pairs: and , and , and
. In other words, these theories cannot establish
infinitely often circuit upper bounds for the corresponding problems. This is
of interest because the weaker theory already formalizes
sophisticated arguments, such as a proof of the PCP Theorem. These consistency
statements are unconditional and improve on earlier theorems of [KO17] and
[BM18] on the consistency of lower bounds with
An average-case depth hierarchy theorem for Boolean circuits
We prove an average-case depth hierarchy theorem for Boolean circuits over
the standard basis of , , and gates.
Our hierarchy theorem says that for every , there is an explicit
-variable Boolean function , computed by a linear-size depth- formula,
which is such that any depth- circuit that agrees with on fraction of all inputs must have size This
answers an open question posed by H{\aa}stad in his Ph.D. thesis.
Our average-case depth hierarchy theorem implies that the polynomial
hierarchy is infinite relative to a random oracle with probability 1,
confirming a conjecture of H{\aa}stad, Cai, and Babai. We also use our result
to show that there is no "approximate converse" to the results of Linial,
Mansour, Nisan and Boppana on the total influence of small-depth circuits, thus
answering a question posed by O'Donnell, Kalai, and Hatami.
A key ingredient in our proof is a notion of \emph{random projections} which
generalize random restrictions
Oracles Are Subtle But Not Malicious
Theoretical computer scientists have been debating the role of oracles since
the 1970's. This paper illustrates both that oracles can give us nontrivial
insights about the barrier problems in circuit complexity, and that they need
not prevent us from trying to solve those problems.
First, we give an oracle relative to which PP has linear-sized circuits, by
proving a new lower bound for perceptrons and low- degree threshold
polynomials. This oracle settles a longstanding open question, and generalizes
earlier results due to Beigel and to Buhrman, Fortnow, and Thierauf. More
importantly, it implies the first nonrelativizing separation of "traditional"
complexity classes, as opposed to interactive proof classes such as MIP and
MA-EXP. For Vinodchandran showed, by a nonrelativizing argument, that PP does
not have circuits of size n^k for any fixed k. We present an alternative proof
of this fact, which shows that PP does not even have quantum circuits of size
n^k with quantum advice. To our knowledge, this is the first nontrivial lower
bound on quantum circuit size.
Second, we study a beautiful algorithm of Bshouty et al. for learning Boolean
circuits in ZPP^NP. We show that the NP queries in this algorithm cannot be
parallelized by any relativizing technique, by giving an oracle relative to
which ZPP^||NP and even BPP^||NP have linear-size circuits. On the other hand,
we also show that the NP queries could be parallelized if P=NP. Thus, classes
such as ZPP^||NP inhabit a "twilight zone," where we need to distinguish
between relativizing and black-box techniques. Our results on this subject have
implications for computational learning theory as well as for the circuit
minimization problem.Comment: 20 pages, 1 figur
Immunity and Simplicity for Exact Counting and Other Counting Classes
Ko [RAIRO 24, 1990] and Bruschi [TCS 102, 1992] showed that in some
relativized world, PSPACE (in fact, ParityP) contains a set that is immune to
the polynomial hierarchy (PH). In this paper, we study and settle the question
of (relativized) separations with immunity for PH and the counting classes PP,
C_{=}P, and ParityP in all possible pairwise combinations. Our main result is
that there is an oracle A relative to which C_{=}P contains a set that is
immune to BPP^{ParityP}. In particular, this C_{=}P^A set is immune to PH^{A}
and ParityP^{A}. Strengthening results of Tor\'{a}n [J.ACM 38, 1991] and Green
[IPL 37, 1991], we also show that, in suitable relativizations, NP contains a
C_{=}P-immune set, and ParityP contains a PP^{PH}-immune set. This implies the
existence of a C_{=}P^{B}-simple set for some oracle B, which extends results
of Balc\'{a}zar et al. [SIAM J.Comp. 14, 1985; RAIRO 22, 1988] and provides the
first example of a simple set in a class not known to be contained in PH. Our
proof technique requires a circuit lower bound for ``exact counting'' that is
derived from Razborov's [Mat. Zametki 41, 1987] lower bound for majority.Comment: 20 page
Towards a complexity theory for the congested clique
The congested clique model of distributed computing has been receiving
attention as a model for densely connected distributed systems. While there has
been significant progress on the side of upper bounds, we have very little in
terms of lower bounds for the congested clique; indeed, it is now know that
proving explicit congested clique lower bounds is as difficult as proving
circuit lower bounds.
In this work, we use various more traditional complexity-theoretic tools to
build a clearer picture of the complexity landscape of the congested clique:
-- Nondeterminism and beyond: We introduce the nondeterministic congested
clique model (analogous to NP) and show that there is a natural canonical
problem family that captures all problems solvable in constant time with
nondeterministic algorithms. We further generalise these notions by introducing
the constant-round decision hierarchy (analogous to the polynomial hierarchy).
-- Non-constructive lower bounds: We lift the prior non-uniform counting
arguments to a general technique for proving non-constructive uniform lower
bounds for the congested clique. In particular, we prove a time hierarchy
theorem for the congested clique, showing that there are decision problems of
essentially all complexities, both in the deterministic and nondeterministic
settings.
-- Fine-grained complexity: We map out relationships between various natural
problems in the congested clique model, arguing that a reduction-based
complexity theory currently gives us a fairly good picture of the complexity
landscape of the congested clique
Logic Meets Algebra: the Case of Regular Languages
The study of finite automata and regular languages is a privileged meeting
point of algebra and logic. Since the work of Buchi, regular languages have
been classified according to their descriptive complexity, i.e. the type of
logical formalism required to define them. The algebraic point of view on
automata is an essential complement of this classification: by providing
alternative, algebraic characterizations for the classes, it often yields the
only opportunity for the design of algorithms that decide expressibility in
some logical fragment.
We survey the existing results relating the expressibility of regular
languages in logical fragments of MSO[S] with algebraic properties of their
minimal automata. In particular, we show that many of the best known results in
this area share the same underlying mechanics and rely on a very strong
relation between logical substitutions and block-products of pseudovarieties of
monoid. We also explain the impact of these connections on circuit complexity
theory.Comment: 37 page
Bounds on the Power of Constant-Depth Quantum Circuits
We show that if a language is recognized within certain error bounds by
constant-depth quantum circuits over a finite family of gates, then it is
computable in (classical) polynomial time. In particular, our results imply
EQNC^0 is contained in P, where EQNC^0 is the constant-depth analog of the
class EQP. On the other hand, we adapt and extend ideas of Terhal and
DiVincenzo (quant-ph/0205133) to show that, for any family F of quantum gates
including Hadamard and CNOT gates, computing the acceptance probabilities of
depth-five circuits over F is just as hard as computing these probabilities for
circuits over F. In particular, this implies that NQNC^0 = NQACC = NQP = coC=P
where NQNC^0 is the constant-depth analog of the class NQP. This essentially
refutes a conjecture of Green et al. that NQACC is contained in TC^0
(quant-ph/0106017)
- …