36,359 research outputs found
Exponential Time Complexity of Weighted Counting of Independent Sets
We consider weighted counting of independent sets using a rational weight x:
Given a graph with n vertices, count its independent sets such that each set of
size k contributes x^k. This is equivalent to computation of the partition
function of the lattice gas with hard-core self-repulsion and hard-core pair
interaction. We show the following conditional lower bounds: If counting the
satisfying assignments of a 3-CNF formula in n variables (#3SAT) needs time
2^{\Omega(n)} (i.e. there is a c>0 such that no algorithm can solve #3SAT in
time 2^{cn}), counting the independent sets of size n/3 of an n-vertex graph
needs time 2^{\Omega(n)} and weighted counting of independent sets needs time
2^{\Omega(n/log^3 n)} for all rational weights x\neq 0.
We have two technical ingredients: The first is a reduction from 3SAT to
independent sets that preserves the number of solutions and increases the
instance size only by a constant factor. Second, we devise a combination of
vertex cloning and path addition. This graph transformation allows us to adapt
a recent technique by Dell, Husfeldt, and Wahlen which enables interpolation by
a family of reductions, each of which increases the instance size only
polylogarithmically.Comment: Introduction revised, differences between versions of counting
independent sets stated more precisely, minor improvements. 14 page
Fine-grained dichotomies for the Tutte plane and Boolean #CSP
Jaeger, Vertigan, and Welsh [15] proved a dichotomy for the complexity of
evaluating the Tutte polynomial at fixed points: The evaluation is #P-hard
almost everywhere, and the remaining points admit polynomial-time algorithms.
Dell, Husfeldt, and Wahl\'en [9] and Husfeldt and Taslaman [12], in combination
with Curticapean [7], extended the #P-hardness results to tight lower bounds
under the counting exponential time hypothesis #ETH, with the exception of the
line , which was left open. We complete the dichotomy theorem for the
Tutte polynomial under #ETH by proving that the number of all acyclic subgraphs
of a given -vertex graph cannot be determined in time unless
#ETH fails.
Another dichotomy theorem we strengthen is the one of Creignou and Hermann
[6] for counting the number of satisfying assignments to a constraint
satisfaction problem instance over the Boolean domain. We prove that all
#P-hard cases are also hard under #ETH. The main ingredient is to prove that
the number of independent sets in bipartite graphs with vertices cannot be
computed in time unless #ETH fails. In order to prove our results,
we use the block interpolation idea by Curticapean [7] and transfer it to
systems of linear equations that might not directly correspond to
interpolation.Comment: 16 pages, 1 figur
Model Counting of Query Expressions: Limitations of Propositional Methods
Query evaluation in tuple-independent probabilistic databases is the problem
of computing the probability of an answer to a query given independent
probabilities of the individual tuples in a database instance. There are two
main approaches to this problem: (1) in `grounded inference' one first obtains
the lineage for the query and database instance as a Boolean formula, then
performs weighted model counting on the lineage (i.e., computes the probability
of the lineage given probabilities of its independent Boolean variables); (2)
in methods known as `lifted inference' or `extensional query evaluation', one
exploits the high-level structure of the query as a first-order formula.
Although it is widely believed that lifted inference is strictly more powerful
than grounded inference on the lineage alone, no formal separation has
previously been shown for query evaluation. In this paper we show such a formal
separation for the first time.
We exhibit a class of queries for which model counting can be done in
polynomial time using extensional query evaluation, whereas the algorithms used
in state-of-the-art exact model counters on their lineages provably require
exponential time. Our lower bounds on the running times of these exact model
counters follow from new exponential size lower bounds on the kinds of d-DNNF
representations of the lineages that these model counters (either explicitly or
implicitly) produce. Though some of these queries have been studied before, no
non-trivial lower bounds on the sizes of these representations for these
queries were previously known.Comment: To appear in International Conference on Database Theory (ICDT) 201
Exponential Time Complexity of the Permanent and the Tutte Polynomial
We show conditional lower bounds for well-studied #P-hard problems:
(a) The number of satisfying assignments of a 2-CNF formula with n variables
cannot be counted in time exp(o(n)), and the same is true for computing the
number of all independent sets in an n-vertex graph.
(b) The permanent of an n x n matrix with entries 0 and 1 cannot be computed
in time exp(o(n)).
(c) The Tutte polynomial of an n-vertex multigraph cannot be computed in time
exp(o(n)) at most evaluation points (x,y) in the case of multigraphs, and it
cannot be computed in time exp(o(n/polylog n)) in the case of simple graphs.
Our lower bounds are relative to (variants of) the Exponential Time
Hypothesis (ETH), which says that the satisfiability of n-variable 3-CNF
formulas cannot be decided in time exp(o(n)). We relax this hypothesis by
introducing its counting version #ETH, namely that the satisfying assignments
cannot be counted in time exp(o(n)). In order to use #ETH for our lower bounds,
we transfer the sparsification lemma for d-CNF formulas to the counting
setting
Solving weighted and counting variants of connectivity problems parameterized by treewidth deterministically in single exponential time
It is well known that many local graph problems, like Vertex Cover and
Dominating Set, can be solved in 2^{O(tw)}|V|^{O(1)} time for graphs G=(V,E)
with a given tree decomposition of width tw. However, for nonlocal problems,
like the fundamental class of connectivity problems, for a long time we did not
know how to do this faster than tw^{O(tw)}|V|^{O(1)}. Recently, Cygan et al.
(FOCS 2011) presented Monte Carlo algorithms for a wide range of connectivity
problems running in time $c^{tw}|V|^{O(1)} for a small constant c, e.g., for
Hamiltonian Cycle and Steiner tree. Naturally, this raises the question whether
randomization is necessary to achieve this runtime; furthermore, it is
desirable to also solve counting and weighted versions (the latter without
incurring a pseudo-polynomial cost in terms of the weights).
We present two new approaches rooted in linear algebra, based on matrix rank
and determinants, which provide deterministic c^{tw}|V|^{O(1)} time algorithms,
also for weighted and counting versions. For example, in this time we can solve
the traveling salesman problem or count the number of Hamiltonian cycles. The
rank-based ideas provide a rather general approach for speeding up even
straightforward dynamic programming formulations by identifying "small" sets of
representative partial solutions; we focus on the case of expressing
connectivity via sets of partitions, but the essential ideas should have
further applications. The determinant-based approach uses the matrix tree
theorem for deriving closed formulas for counting versions of connectivity
problems; we show how to evaluate those formulas via dynamic programming.Comment: 36 page
Large induced subgraphs via triangulations and CMSO
We obtain an algorithmic meta-theorem for the following optimization problem.
Let \phi\ be a Counting Monadic Second Order Logic (CMSO) formula and t be an
integer. For a given graph G, the task is to maximize |X| subject to the
following: there is a set of vertices F of G, containing X, such that the
subgraph G[F] induced by F is of treewidth at most t, and structure (G[F],X)
models \phi.
Some special cases of this optimization problem are the following generic
examples. Each of these cases contains various problems as a special subcase:
1) "Maximum induced subgraph with at most l copies of cycles of length 0
modulo m", where for fixed nonnegative integers m and l, the task is to find a
maximum induced subgraph of a given graph with at most l vertex-disjoint cycles
of length 0 modulo m.
2) "Minimum \Gamma-deletion", where for a fixed finite set of graphs \Gamma\
containing a planar graph, the task is to find a maximum induced subgraph of a
given graph containing no graph from \Gamma\ as a minor.
3) "Independent \Pi-packing", where for a fixed finite set of connected
graphs \Pi, the task is to find an induced subgraph G[F] of a given graph G
with the maximum number of connected components, such that each connected
component of G[F] is isomorphic to some graph from \Pi.
We give an algorithm solving the optimization problem on an n-vertex graph G
in time O(#pmc n^{t+4} f(t,\phi)), where #pmc is the number of all potential
maximal cliques in G and f is a function depending of t and \phi\ only. We also
show how a similar running time can be obtained for the weighted version of the
problem. Pipelined with known bounds on the number of potential maximal
cliques, we deduce that our optimization problem can be solved in time
O(1.7347^n) for arbitrary graphs, and in polynomial time for graph classes with
polynomial number of minimal separators
Structurally Parameterized d-Scattered Set
In -Scattered Set we are given an (edge-weighted) graph and are asked to
select at least vertices, so that the distance between any pair is at least
, thus generalizing Independent Set. We provide upper and lower bounds on
the complexity of this problem with respect to various standard graph
parameters. In particular, we show the following:
- For any , an -time algorithm, where
is the treewidth of the input graph.
- A tight SETH-based lower bound matching this algorithm's performance. These
generalize known results for Independent Set.
- -Scattered Set is W[1]-hard parameterized by vertex cover (for
edge-weighted graphs), or feedback vertex set (for unweighted graphs), even if
is an additional parameter.
- A single-exponential algorithm parameterized by vertex cover for unweighted
graphs, complementing the above-mentioned hardness.
- A -time algorithm parameterized by tree-depth
(), as well as a matching ETH-based lower bound, both for
unweighted graphs.
We complement these mostly negative results by providing an FPT approximation
scheme parameterized by treewidth. In particular, we give an algorithm which,
for any error parameter , runs in time
and returns a
-scattered set of size , if a -scattered set of the same
size exists
A Simple FPTAS for Counting Edge Covers
An edge cover of a graph is a set of edges such that every vertex has at
least an adjacent edge in it. Previously, approximation algorithm for counting
edge covers is only known for 3 regular graphs and it is randomized. We design
a very simple deterministic fully polynomial-time approximation scheme (FPTAS)
for counting the number of edge covers for any graph. Our main technique is
correlation decay, which is a powerful tool to design FPTAS for counting
problems. In order to get FPTAS for general graphs without degree bound, we
make use of a stronger notion called computationally efficient correlation
decay, which is introduced in [Li, Lu, Yin SODA 2012].Comment: To appear in SODA 201
- âŠ