723 research outputs found
Understanding the Complexity of Lifted Inference and Asymmetric Weighted Model Counting
In this paper we study lifted inference for the Weighted First-Order Model
Counting problem (WFOMC), which counts the assignments that satisfy a given
sentence in first-order logic (FOL); it has applications in Statistical
Relational Learning (SRL) and Probabilistic Databases (PDB). We present several
results. First, we describe a lifted inference algorithm that generalizes prior
approaches in SRL and PDB. Second, we provide a novel dichotomy result for a
non-trivial fragment of FO CNF sentences, showing that for each sentence the
WFOMC problem is either in PTIME or #P-hard in the size of the input domain; we
prove that, in the first case our algorithm solves the WFOMC problem in PTIME,
and in the second case it fails. Third, we present several properties of the
algorithm. Finally, we discuss limitations of lifted inference for symmetric
probabilistic databases (where the weights of ground literals depend only on
the relation name, and not on the constants of the domain), and prove the
impossibility of a dichotomy result for the complexity of probabilistic
inference for the entire language FOL
Hypergraph Acyclicity and Propositional Model Counting
We show that the propositional model counting problem #SAT for CNF- formulas
with hypergraphs that allow a disjoint branches decomposition can be solved in
polynomial time. We show that this class of hypergraphs is incomparable to
hypergraphs of bounded incidence cliquewidth which were the biggest class of
hypergraphs for which #SAT was known to be solvable in polynomial time so far.
Furthermore, we present a polynomial time algorithm that computes a disjoint
branches decomposition of a given hypergraph if it exists and rejects
otherwise. Finally, we show that some slight extensions of the class of
hypergraphs with disjoint branches decompositions lead to intractable #SAT,
leaving open how to generalize the counting result of this paper
Achieving New Upper Bounds for the Hypergraph Duality Problem through Logic
The hypergraph duality problem DUAL is defined as follows: given two simple
hypergraphs and , decide whether
consists precisely of all minimal transversals of (in which case
we say that is the dual of ). This problem is
equivalent to deciding whether two given non-redundant monotone DNFs are dual.
It is known that non-DUAL, the complementary problem to DUAL, is in
, where
denotes the complexity class of all problems that after a nondeterministic
guess of bits can be decided (checked) within complexity class
. It was conjectured that non-DUAL is in . In this paper we prove this conjecture and actually
place the non-DUAL problem into the complexity class which is a subclass of . We here refer to the logtime-uniform version of
, which corresponds to , i.e., first order
logic augmented by counting quantifiers. We achieve the latter bound in two
steps. First, based on existing problem decomposition methods, we develop a new
nondeterministic algorithm for non-DUAL that requires to guess
bits. We then proceed by a logical analysis of this algorithm, allowing us to
formulate its deterministic part in . From this result, by
the well known inclusion , it follows
that DUAL belongs also to . Finally, by exploiting
the principles on which the proposed nondeterministic algorithm is based, we
devise a deterministic algorithm that, given two hypergraphs and
, computes in quadratic logspace a transversal of
missing in .Comment: Restructured the presentation in order to be the extended version of
a paper that will shortly appear in SIAM Journal on Computin
Breaking the PPSZ Barrier for Unique 3-SAT
The PPSZ algorithm by Paturi, Pudl\'ak, Saks, and Zane (FOCS 1998) is the
fastest known algorithm for (Promise) Unique k-SAT. We give an improved
algorithm with exponentially faster bounds for Unique 3-SAT.
For uniquely satisfiable 3-CNF formulas, we do the following case
distinction: We call a clause critical if exactly one literal is satisfied by
the unique satisfying assignment. If a formula has many critical clauses, we
observe that PPSZ by itself is already faster. If there are only few clauses
allover, we use an algorithm by Wahlstr\"om (ESA 2005) that is faster than PPSZ
in this case. Otherwise we have a formula with few critical and many
non-critical clauses. Non-critical clauses have at least two literals
satisfied; we show how to exploit this to improve PPSZ.Comment: 13 pages; major revision with simplified algorithm but slightly worse
constant
The output distribution of important LULU-operators
Two procedures to compute the output distribution phi_S of certain stack
filters S (so called erosion-dilation cascades) are given. One rests on the
disjunctive normal form of S and also yields the rank selection probabilities.
The other is based on inclusion-exclusion and e.g. yields phi_S for some
important LULU-operators S. Properties of phi_S can be used to characterize
smoothing properties of S. One of the methods discussed also allows for the
calculation of the reliability polynomial of any positive Boolean function
(e.g. one derived from a connected graph).Comment: 20 pages, up to trivial differences this is the final version to be
published in Quaestiones Mathematicae 201
Counting Complexity for Reasoning in Abstract Argumentation
In this paper, we consider counting and projected model counting of
extensions in abstract argumentation for various semantics. When asking for
projected counts we are interested in counting the number of extensions of a
given argumentation framework while multiple extensions that are identical when
restricted to the projected arguments count as only one projected extension. We
establish classical complexity results and parameterized complexity results
when the problems are parameterized by treewidth of the undirected
argumentation graph. To obtain upper bounds for counting projected extensions,
we introduce novel algorithms that exploit small treewidth of the undirected
argumentation graph of the input instance by dynamic programming (DP). Our
algorithms run in time double or triple exponential in the treewidth depending
on the considered semantics. Finally, we take the exponential time hypothesis
(ETH) into account and establish lower bounds of bounded treewidth algorithms
for counting extensions and projected extension.Comment: Extended version of a paper published at AAAI-1
- …