18,070 research outputs found
Distributed local approximation algorithms for maximum matching in graphs and hypergraphs
We describe approximation algorithms in Linial's classic LOCAL model of
distributed computing to find maximum-weight matchings in a hypergraph of rank
. Our main result is a deterministic algorithm to generate a matching which
is an -approximation to the maximum weight matching, running in rounds. (Here, the
notations hides and factors).
This is based on a number of new derandomization techniques extending methods
of Ghaffari, Harris & Kuhn (2017).
As a main application, we obtain nearly-optimal algorithms for the
long-studied problem of maximum-weight graph matching. Specifically, we get a
approximation algorithm using randomized time and deterministic time.
The second application is a faster algorithm for hypergraph maximal matching,
a versatile subroutine introduced in Ghaffari et al. (2017) for a variety of
local graph algorithms. This gives an algorithm for -edge-list
coloring in rounds deterministically or
rounds randomly. Another consequence (with
additional optimizations) is an algorithm which generates an edge-orientation
with out-degree at most for a graph of
arboricity ; for fixed this runs in
rounds deterministically or rounds randomly
Some results on chromatic number as a function of triangle count
A variety of powerful extremal results have been shown for the chromatic
number of triangle-free graphs. Three noteworthy bounds are in terms of the
number of vertices, edges, and maximum degree given by Poljak \& Tuza (1994),
and Johansson. There have been comparatively fewer works extending these types
of bounds to graphs with a small number of triangles. One noteworthy exception
is a result of Alon et. al (1999) bounding the chromatic number for graphs with
low degree and few triangles per vertex; this bound is nearly the same as for
triangle-free graphs. This type of parametrization is much less rigid, and has
appeared in dozens of combinatorial constructions.
In this paper, we show a similar type of result for as a function
of the number of vertices , the number of edges , as well as the triangle
count (both local and global measures). Our results smoothly interpolate
between the generic bounds true for all graphs and bounds for triangle-free
graphs. Our results are tight for most of these cases; we show how an open
problem regarding fractional chromatic number and degeneracy in triangle-free
graphs can resolve the small remaining gap in our bounds
Deterministic parallel algorithms for bilinear objective functions
Many randomized algorithms can be derandomized efficiently using either the
method of conditional expectations or probability spaces with low independence.
A series of papers, beginning with work by Luby (1988), showed that in many
cases these techniques can be combined to give deterministic parallel (NC)
algorithms for a variety of combinatorial optimization problems, with low time-
and processor-complexity.
We extend and generalize a technique of Luby for efficiently handling
bilinear objective functions. One noteworthy application is an NC algorithm for
maximal independent set. On a graph with edges and vertices, this
takes time and processors, nearly
matching the best randomized parallel algorithms. Other applications include
reduced processor counts for algorithms of Berger (1997) for maximum acyclic
subgraph and Gale-Berlekamp switching games.
This bilinear factorization also gives better algorithms for problems
involving discrepancy. An important application of this is to automata-fooling
probability spaces, which are the basis of a notable derandomization technique
of Sivakumar (2002). Our method leads to large reduction in processor
complexity for a number of derandomization algorithms based on
automata-fooling, including set discrepancy and the Johnson-Lindenstrauss
Lemma
Parallel algorithms and concentration bounds for the Lovasz Local Lemma via witness DAGs
The Lov\'{a}sz Local Lemma (LLL) is a cornerstone principle in the
probabilistic method of combinatorics, and a seminal algorithm of Moser &
Tardos (2010) provides an efficient randomized algorithm to implement it. This
can be parallelized to give an algorithm that uses polynomially many processors
and runs in time on an EREW PRAM, stemming from
adaptive computations of a maximal independent set (MIS). Chung et al. (2014)
developed faster local and parallel algorithms, potentially running in time
, but these algorithms require more stringent conditions than the
LLL.
We give a new parallel algorithm that works under essentially the same
conditions as the original algorithm of Moser & Tardos but uses only a single
MIS computation, thus running in time on an EREW PRAM. This can
be derandomized to give an NC algorithm running in time as well,
speeding up a previous NC LLL algorithm of Chandrasekaran et al. (2013).
We also provide improved and tighter bounds on the run-times of the
sequential and parallel resampling-based algorithms originally developed by
Moser & Tardos. These apply to any problem instance in which the tighter
Shearer LLL criterion is satisfied
Tight bounds and conjectures for the isolation lemma
Given a hypergraph and a weight function on its vertices, we say that is isolating if there is exactly one edge
of minimum weight . The Isolation Lemma is a
combinatorial principle introduced in Mulmuley et. al (1987) which gives a
lower bound on the number of isolating weight functions. Mulmuley used this as
the basis of a parallel algorithm for finding perfect graph matchings. It has a
number of other applications to parallel algorithms and to reductions of
general search problems to unique search problems (in which there are one or
zero solutions).
The original bound given by Mulmuley et al. was recently improved by Ta-Shma
(2015). In this paper, we show improved lower bounds on the number of isolating
weight functions, and we conjecture that the extremal case is when consists
of singleton edges. When our improved bound matches this extremal
case asymptotically.
We are able to show that this conjecture holds in a number of special cases:
when is a linear hypergraph or is 1-degenerate, or when . We also
show that it holds asymptotically when
The Moser-Tardos Framework with Partial Resampling
The resampling algorithm of Moser \& Tardos is a powerful approach to develop
constructive versions of the Lov\'{a}sz Local Lemma (LLL). We generalize this
to partial resampling: when a bad event holds, we resample an
appropriately-random subset of the variables that define this event, rather
than the entire set as in Moser & Tardos. This is particularly useful when the
bad events are determined by sums of random variables. This leads to several
improved algorithmic applications in scheduling, graph transversals, packet
routing etc. For instance, we settle a conjecture of Szab\'{o} & Tardos (2006)
on graph transversals asymptotically, and obtain improved approximation ratios
for a packet routing problem of Leighton, Maggs, & Rao (1994)
Improved bounds and algorithms for graph cuts and network reliability
Karger (SIAM Journal on Computing, 1999) developed the first fully-polynomial
approximation scheme to estimate the probability that a graph becomes
disconnected, given that its edges are removed independently with probability
. This algorithm runs in time to obtain an
estimate within relative error .
We improve this run-time through algorithmic and graph-theoretic advances.
First, there is a certain key sub-problem encountered by Karger, for which a
generic estimation procedure is employed, we show that this has a special
structure for which a much more efficient algorithm can be used. Second, we
show better bounds on the number of edge cuts which are likely to fail. Here,
Karger's analysis uses a variety of bounds for various graph parameters, we
show that these bounds cannot be simultaneously tight. We describe a new graph
parameter, which simultaneously influences all the bounds used by Karger, and
obtain much tighter estimates of the cut structure of . These techniques
allow us to improve the runtime to , our results also
rigorously prove certain experimental observations of Karger & Tai (Proc.
ACM-SIAM Symposium on Discrete Algorithms, 1997). Our rigorous proofs are
motivated by certain non-rigorous differential-equation approximations which,
however, provably track the worst-case trajectories of the relevant parameters.
A key driver of Karger's approach (and other cut-related results) is a bound
on the number of small cuts: we improve these estimates when the min-cut size
is "small" and odd, augmenting, in part, a result of Bixby (Bulletin of the
AMS, 1974)
Parameter estimation for integer-valued Gibbs distributions
We consider Gibbs distributions, which are families of probability
distributions over a discrete space with probability mass function
given by . Here
is a fixed function (called a Hamiltonian),
is the parameter of the distribution, and the normalization factor
is called
the partition function. We study how function can be estimated using an
oracle that produces samples for a value in
a given interval .
We consider the problem of estimating the normalized coefficients for
indices satisfying
, where is a
given parameter and is a given subset of . We solve this using
samples where
, and we show this is optimal up
to logarithmic factors. We also improve the sample complexity to roughly
for applications where
the coefficients are log-concave (e.g. counting connected subgraphs of a given
graph).
As a key subroutine, we show how to estimate using samples. This improves over a prior
algorithm of Kolmogorov (2018) that uses
samples. We also show a "batched" version of this algorithm which
simultaneously estimates for many values of
, at essentially the same cost as for estimating just
alone. We show matching lower bounds,
demonstrating that this complexity is optimal as a function of up to
logarithmic terms.Comment: Superseded by arXiv:2007.1082
Algorithmic and enumerative aspects of the Moser-Tardos distribution
Moser & Tardos have developed a powerful algorithmic approach (henceforth
"MT") to the Lovasz Local Lemma (LLL); the basic operation done in MT and its
variants is a search for "bad" events in a current configuration. In the
initial stage of MT, the variables are set independently. We examine the
distributions on these variables which arise during intermediate stages of MT.
We show that these configurations have a more or less "random" form, building
further on the "MT-distribution" concept of Haeupler et al. in understanding
the (intermediate and) output distribution of MT. This has a variety of
algorithmic applications; the most important is that bad events can be found
relatively quickly, improving upon MT across the complexity spectrum: it makes
some polynomial-time algorithms sub-linear (e.g., for Latin transversals, which
are of basic combinatorial interest), gives lower-degree polynomial run-times
in some settings, transforms certain super-polynomial-time algorithms into
polynomial-time ones, and leads to Las Vegas algorithms for some coloring
problems for which only Monte Carlo algorithms were known.
We show that in certain conditions when the LLL condition is violated, a
variant of the MT algorithm can still produce a distribution which avoids most
of the bad events. We show in some cases this MT variant can run faster than
the original MT algorithm itself, and develop the first-known criterion for the
case of the asymmetric LLL. This can be used to find partial Latin transversals
-- improving upon earlier bounds of Stein (1975) -- among other applications.
We furthermore give applications in enumeration, showing that most applications
(where we aim for all or most of the bad events to be avoided) have many more
solutions than known before by proving that the MT-distribution has "large"
min-entropy and hence that its support-size is large
- …