17,938 research outputs found
Lower Bounds on Quantum Query Complexity
Shor's and Grover's famous quantum algorithms for factoring and searching
show that quantum computers can solve certain computational problems
significantly faster than any classical computer. We discuss here what quantum
computers_cannot_ do, and specifically how to prove limits on their
computational power. We cover the main known techniques for proving lower
bounds, and exemplify and compare the methods.Comment: survey, 23 page
Approximate resilience, monotonicity, and the complexity of agnostic learning
A function is -resilient if all its Fourier coefficients of degree at
most are zero, i.e., is uncorrelated with all low-degree parities. We
study the notion of of Boolean
functions, where we say that is -approximately -resilient if
is -close to a -valued -resilient function in
distance. We show that approximate resilience essentially characterizes the
complexity of agnostic learning of a concept class over the uniform
distribution. Roughly speaking, if all functions in a class are far from
being -resilient then can be learned agnostically in time and
conversely, if contains a function close to being -resilient then
agnostic learning of in the statistical query (SQ) framework of Kearns has
complexity of at least . This characterization is based on the
duality between approximation by degree- polynomials and
approximate -resilience that we establish. In particular, it implies that
approximation by low-degree polynomials, known to be sufficient for
agnostic learning over product distributions, is in fact necessary.
Focusing on monotone Boolean functions, we exhibit the existence of
near-optimal -approximately
-resilient monotone functions for all
. Prior to our work, it was conceivable even that every monotone
function is -far from any -resilient function. Furthermore, we
construct simple, explicit monotone functions based on and that are close to highly resilient functions. Our constructions are
based on a fairly general resilience analysis and amplification. These
structural results, together with the characterization, imply nearly optimal
lower bounds for agnostic learning of monotone juntas
Fourier sparsity, spectral norm, and the Log-rank conjecture
We study Boolean functions with sparse Fourier coefficients or small spectral
norm, and show their applications to the Log-rank Conjecture for XOR functions
f(x\oplus y) --- a fairly large class of functions including well studied ones
such as Equality and Hamming Distance. The rank of the communication matrix M_f
for such functions is exactly the Fourier sparsity of f. Let d be the F2-degree
of f and D^CC(f) stand for the deterministic communication complexity for
f(x\oplus y). We show that 1. D^CC(f) = O(2^{d^2/2} log^{d-2} ||\hat f||_1). In
particular, the Log-rank conjecture holds for XOR functions with constant
F2-degree. 2. D^CC(f) = O(d ||\hat f||_1) = O(\sqrt{rank(M_f)}\logrank(M_f)).
We obtain our results through a degree-reduction protocol based on a variant of
polynomial rank, and actually conjecture that its communication cost is already
\log^{O(1)}rank(M_f). The above bounds also hold for the parity decision tree
complexity of f, a measure that is no less than the communication complexity
(up to a factor of 2).
Along the way we also show several structural results about Boolean functions
with small F2-degree or small spectral norm, which could be of independent
interest. For functions f with constant F2-degree: 1) f can be written as the
summation of quasi-polynomially many indicator functions of subspaces with
\pm-signs, improving the previous doubly exponential upper bound by Green and
Sanders; 2) being sparse in Fourier domain is polynomially equivalent to having
a small parity decision tree complexity; 3) f depends only on polylog||\hat
f||_1 linear functions of input variables. For functions f with small spectral
norm: 1) there is an affine subspace with co-dimension O(||\hat f||_1) on which
f is a constant; 2) there is a parity decision tree with depth O(||\hat f||_1
log ||\hat f||_0).Comment: v2: Corollary 31 of v1 removed because of a bug in the proof. (Other
results not affected.
Low-Sensitivity Functions from Unambiguous Certificates
We provide new query complexity separations against sensitivity for total
Boolean functions: a power separation between deterministic (and even
randomized or quantum) query complexity and sensitivity, and a power
separation between certificate complexity and sensitivity. We get these
separations by using a new connection between sensitivity and a seemingly
unrelated measure called one-sided unambiguous certificate complexity
(). We also show that is lower-bounded by fractional block
sensitivity, which means we cannot use these techniques to get a
super-quadratic separation between and . We also provide a
quadratic separation between the tree-sensitivity and decision tree complexity
of Boolean functions, disproving a conjecture of Gopalan, Servedio, Tal, and
Wigderson (CCC 2016).
Along the way, we give a power separation between certificate
complexity and one-sided unambiguous certificate complexity, improving the
power separation due to G\"o\"os (FOCS 2015). As a consequence, we
obtain an improved lower-bound on the
co-nondeterministic communication complexity of the Clique vs. Independent Set
problem.Comment: 25 pages. This version expands the results and adds Pooya Hatami and
Avishay Tal as author
Improving Quantum Query Complexity of Boolean Matrix Multiplication Using Graph Collision
The quantum query complexity of Boolean matrix multiplication is typically
studied as a function of the matrix dimension, n, as well as the number of 1s
in the output, \ell. We prove an upper bound of O (n\sqrt{\ell}) for all values
of \ell. This is an improvement over previous algorithms for all values of
\ell. On the other hand, we show that for any \eps < 1 and any \ell <= \eps
n^2, there is an \Omega(n\sqrt{\ell}) lower bound for this problem, showing
that our algorithm is essentially tight.
We first reduce Boolean matrix multiplication to several instances of graph
collision. We then provide an algorithm that takes advantage of the fact that
the underlying graph in all of our instances is very dense to find all graph
collisions efficiently
- …