4,210 research outputs found
Three Puzzles on Mathematics, Computation, and Games
In this lecture I will talk about three mathematical puzzles involving
mathematics and computation that have preoccupied me over the years. The first
puzzle is to understand the amazing success of the simplex algorithm for linear
programming. The second puzzle is about errors made when votes are counted
during elections. The third puzzle is: are quantum computers possible?Comment: ICM 2018 plenary lecture, Rio de Janeiro, 36 pages, 7 Figure
Sum of squares lower bounds for refuting any CSP
Let be a nontrivial -ary predicate. Consider a
random instance of the constraint satisfaction problem on
variables with constraints, each being applied to randomly
chosen literals. Provided the constraint density satisfies , such
an instance is unsatisfiable with high probability. The \emph{refutation}
problem is to efficiently find a proof of unsatisfiability.
We show that whenever the predicate supports a -\emph{wise uniform}
probability distribution on its satisfying assignments, the sum of squares
(SOS) algorithm of degree
(which runs in time ) \emph{cannot} refute a random instance of
. In particular, the polynomial-time SOS algorithm requires
constraints to refute random instances of
CSP when supports a -wise uniform distribution on its satisfying
assignments. Together with recent work of Lee et al. [LRS15], our result also
implies that \emph{any} polynomial-size semidefinite programming relaxation for
refutation requires at least constraints.
Our results (which also extend with no change to CSPs over larger alphabets)
subsume all previously known lower bounds for semialgebraic refutation of
random CSPs. For every constraint predicate~, they give a three-way hardness
tradeoff between the density of constraints, the SOS degree (hence running
time), and the strength of the refutation. By recent algorithmic results of
Allen et al. [AOW15] and Raghavendra et al. [RRS16], this full three-way
tradeoff is \emph{tight}, up to lower-order factors.Comment: 39 pages, 1 figur
Efficiently Detecting Torsion Points and Subtori
Suppose X is the complex zero set of a finite collection of polynomials in
Z[x_1,...,x_n]. We show that deciding whether X contains a point all of whose
coordinates are d_th roots of unity can be done within NP^NP (relative to the
sparse encoding), under a plausible assumption on primes in arithmetic
progression. In particular, our hypothesis can still hold even under certain
failures of the Generalized Riemann Hypothesis, such as the presence of
Siegel-Landau zeroes. Furthermore, we give a similar (but UNconditional)
complexity upper bound for n=1. Finally, letting T be any algebraic subgroup of
(C^*)^n we show that deciding whether X contains T is coNP-complete (relative
to an even more efficient encoding),unconditionally. We thus obtain new
non-trivial families of multivariate polynomial systems where deciding the
existence of complex roots can be done unconditionally in the polynomial
hierarchy -- a family of complexity classes lying between PSPACE and P,
intimately connected with the P=?NP Problem. We also discuss a connection to
Laurent's solution of Chabauty's Conjecture from arithmetic geometry.Comment: 21 pages, no figures. Final version, with additional commentary and
references. Also fixes a gap in Theorems 2 (now Theorem 1.3) regarding
translated subtor
Complexity Theory, Game Theory, and Economics: The Barbados Lectures
This document collects the lecture notes from my mini-course "Complexity
Theory, Game Theory, and Economics," taught at the Bellairs Research Institute
of McGill University, Holetown, Barbados, February 19--23, 2017, as the 29th
McGill Invitational Workshop on Computational Complexity.
The goal of this mini-course is twofold: (i) to explain how complexity theory
has helped illuminate several barriers in economics and game theory; and (ii)
to illustrate how game-theoretic questions have led to new and interesting
complexity theory, including recent several breakthroughs. It consists of two
five-lecture sequences: the Solar Lectures, focusing on the communication and
computational complexity of computing equilibria; and the Lunar Lectures,
focusing on applications of complexity theory in game theory and economics. No
background in game theory is assumed.Comment: Revised v2 from December 2019 corrects some errors in and adds some
recent citations to v1 Revised v3 corrects a few typos in v
Rounding Sum-of-Squares Relaxations
We present a general approach to rounding semidefinite programming
relaxations obtained by the Sum-of-Squares method (Lasserre hierarchy). Our
approach is based on using the connection between these relaxations and the
Sum-of-Squares proof system to transform a *combining algorithm* -- an
algorithm that maps a distribution over solutions into a (possibly weaker)
solution -- into a *rounding algorithm* that maps a solution of the relaxation
to a solution of the original problem.
Using this approach, we obtain algorithms that yield improved results for
natural variants of three well-known problems:
1) We give a quasipolynomial-time algorithm that approximates the maximum of
a low degree multivariate polynomial with non-negative coefficients over the
Euclidean unit sphere. Beyond being of interest in its own right, this is
related to an open question in quantum information theory, and our techniques
have already led to improved results in this area (Brand\~{a}o and Harrow, STOC
'13).
2) We give a polynomial-time algorithm that, given a d dimensional subspace
of R^n that (almost) contains the characteristic function of a set of size n/k,
finds a vector in the subspace satisfying ,
where . Aside from being a natural relaxation, this
is also motivated by a connection to the Small Set Expansion problem shown by
Barak et al. (STOC 2012) and our results yield a certain improvement for that
problem.
3) We use this notion of L_4 vs. L_2 sparsity to obtain a polynomial-time
algorithm with substantially improved guarantees for recovering a planted
-sparse vector v in a random d-dimensional subspace of R^n. If v has mu n
nonzero coordinates, we can recover it with high probability whenever , improving for prior methods which
intrinsically required
Fast counting with tensor networks
We introduce tensor network contraction algorithms for counting satisfying
assignments of constraint satisfaction problems (#CSPs). We represent each
arbitrary #CSP formula as a tensor network, whose full contraction yields the
number of satisfying assignments of that formula, and use graph theoretical
methods to determine favorable orders of contraction. We employ our heuristics
for the solution of #P-hard counting boolean satisfiability (#SAT) problems,
namely monotone #1-in-3SAT and #Cubic-Vertex-Cover, and find that they
outperform state-of-the-art solvers by a significant margin.Comment: v2: added results for monotone #1-in-3SAT; published versio
Some Applications of Coding Theory in Computational Complexity
Error-correcting codes and related combinatorial constructs play an important
role in several recent (and old) results in computational complexity theory. In
this paper we survey results on locally-testable and locally-decodable
error-correcting codes, and their applications to complexity theory and to
cryptography.
Locally decodable codes are error-correcting codes with sub-linear time
error-correcting algorithms. They are related to private information retrieval
(a type of cryptographic protocol), and they are used in average-case
complexity and to construct ``hard-core predicates'' for one-way permutations.
Locally testable codes are error-correcting codes with sub-linear time
error-detection algorithms, and they are the combinatorial core of
probabilistically checkable proofs
Sum-of-squares lower bounds for planted clique
Finding cliques in random graphs and the closely related "planted" clique
variant, where a clique of size k is planted in a random G(n, 1/2) graph, have
been the focus of substantial study in algorithm design. Despite much effort,
the best known polynomial-time algorithms only solve the problem for k ~
sqrt(n).
In this paper we study the complexity of the planted clique problem under
algorithms from the Sum-of-squares hierarchy. We prove the first average case
lower bound for this model: for almost all graphs in G(n,1/2), r rounds of the
SOS hierarchy cannot find a planted k-clique unless k > n^{1/2r} (up to
logarithmic factors). Thus, for any constant number of rounds planted cliques
of size n^{o(1)} cannot be found by this powerful class of algorithms. This is
shown via an integrability gap for the natural formulation of maximum clique
problem on random graphs for SOS and Lasserre hierarchies, which in turn follow
from degree lower bounds for the Positivestellensatz proof system.
We follow the usual recipe for such proofs. First, we introduce a natural
"dual certificate" (also known as a "vector-solution" or "pseudo-expectation")
for the given system of polynomial equations representing the problem for every
fixed input graph. Then we show that the matrix associated with this dual
certificate is PSD (positive semi-definite) with high probability over the
choice of the input graph.This requires the use of certain tools. One is the
theory of association schemes, and in particular the eigenspaces and
eigenvalues of the Johnson scheme. Another is a combinatorial method we develop
to compute (via traces) norm bounds for certain random matrices whose entries
are highly dependent; we hope this method will be useful elsewhere
- …