1,027 research outputs found
Distributed local approximation algorithms for maximum matching in graphs and hypergraphs
We describe approximation algorithms in Linial's classic LOCAL model of
distributed computing to find maximum-weight matchings in a hypergraph of rank
. Our main result is a deterministic algorithm to generate a matching which
is an -approximation to the maximum weight matching, running in rounds. (Here, the
notations hides and factors).
This is based on a number of new derandomization techniques extending methods
of Ghaffari, Harris & Kuhn (2017).
As a main application, we obtain nearly-optimal algorithms for the
long-studied problem of maximum-weight graph matching. Specifically, we get a
approximation algorithm using randomized time and deterministic time.
The second application is a faster algorithm for hypergraph maximal matching,
a versatile subroutine introduced in Ghaffari et al. (2017) for a variety of
local graph algorithms. This gives an algorithm for -edge-list
coloring in rounds deterministically or
rounds randomly. Another consequence (with
additional optimizations) is an algorithm which generates an edge-orientation
with out-degree at most for a graph of
arboricity ; for fixed this runs in
rounds deterministically or rounds randomly
Improved Pseudorandom Generators from Pseudorandom Multi-Switching Lemmas
We give the best known pseudorandom generators for two touchstone classes in
unconditional derandomization: an -PRG for the class of size-
depth- circuits with seed length , and an -PRG for the class of -sparse
polynomials with seed length . These results bring the state of the art for
unconditional derandomization of these classes into sharp alignment with the
state of the art for computational hardness for all parameter settings:
improving on the seed lengths of either PRG would require breakthrough progress
on longstanding and notorious circuit lower bounds.
The key enabling ingredient in our approach is a new \emph{pseudorandom
multi-switching lemma}. We derandomize recently-developed
\emph{multi}-switching lemmas, which are powerful generalizations of
H{\aa}stad's switching lemma that deal with \emph{families} of depth-two
circuits. Our pseudorandom multi-switching lemma---a randomness-efficient
algorithm for sampling restrictions that simultaneously simplify all circuits
in a family---achieves the parameters obtained by the (full randomness)
multi-switching lemmas of Impagliazzo, Matthews, and Paturi [IMP12] and
H{\aa}stad [H{\aa}s14]. This optimality of our derandomization translates into
the optimality (given current circuit lower bounds) of our PRGs for
and sparse polynomials
On Derandomizing Local Distributed Algorithms
The gap between the known randomized and deterministic local distributed
algorithms underlies arguably the most fundamental and central open question in
distributed graph algorithms. In this paper, we develop a generic and clean
recipe for derandomizing LOCAL algorithms. We also exhibit how this simple
recipe leads to significant improvements on a number of problem. Two main
results are:
- An improved distributed hypergraph maximal matching algorithm, improving on
Fischer, Ghaffari, and Kuhn [FOCS'17], and giving improved algorithms for
edge-coloring, maximum matching approximation, and low out-degree edge
orientation. The first gives an improved algorithm for Open Problem 11.4 of the
book of Barenboim and Elkin, and the last gives the first positive resolution
of their Open Problem 11.10.
- An improved distributed algorithm for the Lov\'{a}sz Local Lemma, which
gets closer to a conjecture of Chang and Pettie [FOCS'17], and moreover leads
to improved distributed algorithms for problems such as defective coloring and
-SAT.Comment: 37 page
Pseudorandomness for Approximate Counting and Sampling
We study computational procedures that use both randomness and nondeterminism. The goal of this paper is to derandomize such procedures under the weakest possible assumptions.
Our main technical contribution allows one to “boost” a given hardness assumption: We show that if there is a problem in EXP that cannot be computed by poly-size nondeterministic circuits then there is one which cannot be computed by poly-size circuits that make non-adaptive NP oracle queries. This in particular shows that the various assumptions used over the last few years by several authors to derandomize Arthur-Merlin games (i.e., show AM = NP) are in fact all equivalent.
We also define two new primitives that we regard as the natural pseudorandom objects associated with approximate counting and sampling of NP-witnesses. We use the “boosting” theorem and hashing techniques to construct these primitives using an assumption that is no stronger than that used to derandomize AM.
We observe that Cai's proof that S_2^P ⊆ PP⊆(NP) and the learning algorithm of Bshouty et al. can be seen as reductions to sampling that are not probabilistic. As a consequence they can be derandomized under an assumption which is weaker than the assumption that was previously known to suffice
An Atypical Survey of Typical-Case Heuristic Algorithms
Heuristic approaches often do so well that they seem to pretty much always
give the right answer. How close can heuristic algorithms get to always giving
the right answer, without inducing seismic complexity-theoretic consequences?
This article first discusses how a series of results by Berman, Buhrman,
Hartmanis, Homer, Longpr\'{e}, Ogiwara, Sch\"{o}ening, and Watanabe, from the
early 1970s through the early 1990s, explicitly or implicitly limited how well
heuristic algorithms can do on NP-hard problems. In particular, many desirable
levels of heuristic success cannot be obtained unless severe, highly unlikely
complexity class collapses occur. Second, we survey work initiated by Goldreich
and Wigderson, who showed how under plausible assumptions deterministic
heuristics for randomized computation can achieve a very high frequency of
correctness. Finally, we consider formal ways in which theory can help explain
the effectiveness of heuristics that solve NP-hard problems in practice.Comment: This article is currently scheduled to appear in the December 2012
issue of SIGACT New
Efficient Approximation Algorithms for Multi-Antennae Largest Weight Data Retrieval
In a mobile network, wireless data broadcast over channels (frequencies)
is a powerful means for distributed dissemination of data to clients who access
the channels through multi-antennae equipped on their mobile devices. The
-antennae largest weight data retrieval (ALWDR) problem is to
compute a schedule for downloading a subset of data items that has a maximum
total weight using antennae in a given time interval. In this paper,
we propose a ratio approximation algorithm for the
-antennae largest weight data retrieval (ALWDR) problem that
has the same ratio as the known result but a significantly improved time
complexity of from
when
\cite{lu2014data}. To our knowledge, our algorithm represents the first ratio
approximation solution to ALWDR for the
general case of arbitrary . To achieve this, we first give a ratio
algorithm for the -separated ALWDR
(ALWDR) with runtime , under the assumption
that every data item appears at most once in each segment of
ALWDR, for any input of maximum length on channels in
time slots. Then, we show that we can retain the same ratio for
ALWDR without this assumption at the cost of increased time
complexity to . This result immediately yields an
approximation solution of same ratio and time complexity for ALWDR,
presenting a significant improvement of the known time complexity of ratio
approximation to the problem
- …