431 research outputs found
Belief Propagation and Loop Series on Planar Graphs
We discuss a generic model of Bayesian inference with binary variables
defined on edges of a planar graph. The Loop Calculus approach of [1, 2] is
used to evaluate the resulting series expansion for the partition function. We
show that, for planar graphs, truncating the series at single-connected loops
reduces, via a map reminiscent of the Fisher transformation [3], to evaluating
the partition function of the dimer matching model on an auxiliary planar
graph. Thus, the truncated series can be easily re-summed, using the Pfaffian
formula of Kasteleyn [4]. This allows to identify a big class of
computationally tractable planar models reducible to a dimer model via the
Belief Propagation (gauge) transformation. The Pfaffian representation can also
be extended to the full Loop Series, in which case the expansion becomes a sum
of Pfaffian contributions, each associated with dimer matchings on an extension
to a subgraph of the original graph. Algorithmic consequences of the Pfaffian
representation, as well as relations to quantum and non-planar models, are
discussed.Comment: Accepted for publication in Journal of Statistical Mechanics: theory
and experimen
Combinatorics
This is the report on the Oberwolfach workshop on Combinatorics, held 1â7 January 2006. Combinatorics is a branch of mathematics studying families of mainly, but not exclusively, ïŹnite or countable structures â discrete objects. The discrete objects considered in the workshop were graphs, set systems, discrete geometries, and matrices. The programme consisted of 15 invited lectures, 18 contributed talks, and a problem session focusing on recent developments in graph theory, coding theory, discrete geometry, extremal combinatorics, Ramsey theory, theoretical computer science, and probabilistic combinatorics
Graphical Models and Symmetries : Loopy Belief Propagation Approaches
Whenever a person or an automated system has to reason in uncertain domains, probability theory is necessary. Probabilistic graphical models allow us to build statistical models that capture complex dependencies between random variables. Inference in these models, however, can easily become intractable. Typical ways to address this scaling issue are inference by approximate message-passing, stochastic gradients, and MapReduce, among others. Exploiting the symmetries of graphical models, however, has not yet been considered for scaling statistical machine learning applications. One instance of graphical models that are inherently symmetric are statistical relational models. These have recently gained attraction within the machine learning and AI communities and combine probability theory with first-order logic, thereby allowing for an efficient representation of structured relational domains. The provided formalisms to compactly represent complex real-world domains enable us to effectively describe large problem instances. Inference within and training of graphical models, however, have not been able to keep pace with the increased representational power. This thesis tackles two major aspects of graphical models and shows that both inference and training can indeed benefit from exploiting symmetries. It first deals with efficient inference exploiting symmetries in graphical models for various query types. We introduce lifted loopy belief propagation (lifted LBP), the first lifted parallel inference approach for relational as well as propositional graphical models. Lifted LBP can effectively speed up marginal inference, but cannot straightforwardly be applied to other types of queries. Thus we also demonstrate efficient lifted algorithms for MAP inference and higher order marginals, as well as the efficient handling of multiple inference tasks. Then we turn to the training of graphical models and introduce the first lifted online training for relational models. Our training procedure and the MapReduce lifting for loopy belief propagation combine lifting with the traditional statistical approaches to scaling, thereby bridging the gap between statistical relational learning and traditional statistical machine learning
A Tight Lower Bound for Counting Hamiltonian Cycles via Matrix Rank
For even , the matchings connectivity matrix encodes which
pairs of perfect matchings on vertices form a single cycle. Cygan et al.
(STOC 2013) showed that the rank of over is
and used this to give an
time algorithm for counting Hamiltonian cycles modulo on graphs of
pathwidth . The same authors complemented their algorithm by an
essentially tight lower bound under the Strong Exponential Time Hypothesis
(SETH). This bound crucially relied on a large permutation submatrix within
, which enabled a "pattern propagation" commonly used in previous
related lower bounds, as initiated by Lokshtanov et al. (SODA 2011).
We present a new technique for a similar pattern propagation when only a
black-box lower bound on the asymptotic rank of is given; no
stronger structural insights such as the existence of large permutation
submatrices in are needed. Given appropriate rank bounds, our
technique yields lower bounds for counting Hamiltonian cycles (also modulo
fixed primes ) parameterized by pathwidth.
To apply this technique, we prove that the rank of over the
rationals is . We also show that the rank of
over is for any prime
and even for some primes.
As a consequence, we obtain that Hamiltonian cycles cannot be counted in time
for any unless SETH fails. This
bound is tight due to a time algorithm by Bodlaender et
al. (ICALP 2013). Under SETH, we also obtain that Hamiltonian cycles cannot be
counted modulo primes in time , indicating
that the modulus can affect the complexity in intricate ways.Comment: improved lower bounds modulo primes, improved figures, to appear in
SODA 201
- âŠ