1,163 research outputs found
Understanding the Complexity of Lifted Inference and Asymmetric Weighted Model Counting
In this paper we study lifted inference for the Weighted First-Order Model
Counting problem (WFOMC), which counts the assignments that satisfy a given
sentence in first-order logic (FOL); it has applications in Statistical
Relational Learning (SRL) and Probabilistic Databases (PDB). We present several
results. First, we describe a lifted inference algorithm that generalizes prior
approaches in SRL and PDB. Second, we provide a novel dichotomy result for a
non-trivial fragment of FO CNF sentences, showing that for each sentence the
WFOMC problem is either in PTIME or #P-hard in the size of the input domain; we
prove that, in the first case our algorithm solves the WFOMC problem in PTIME,
and in the second case it fails. Third, we present several properties of the
algorithm. Finally, we discuss limitations of lifted inference for symmetric
probabilistic databases (where the weights of ground literals depend only on
the relation name, and not on the constants of the domain), and prove the
impossibility of a dichotomy result for the complexity of probabilistic
inference for the entire language FOL
Symmetric Weighted First-Order Model Counting
The FO Model Counting problem (FOMC) is the following: given a sentence
in FO and a number , compute the number of models of over a
domain of size ; the Weighted variant (WFOMC) generalizes the problem by
associating a weight to each tuple and defining the weight of a model to be the
product of weights of its tuples. In this paper we study the complexity of the
symmetric WFOMC, where all tuples of a given relation have the same weight. Our
motivation comes from an important application, inference in Knowledge Bases
with soft constraints, like Markov Logic Networks, but the problem is also of
independent theoretical interest. We study both the data complexity, and the
combined complexity of FOMC and WFOMC. For the data complexity we prove the
existence of an FO formula for which FOMC is #P-complete, and the
existence of a Conjunctive Query for which WFOMC is #P-complete. We also
prove that all -acyclic queries have polynomial time data complexity.
For the combined complexity, we prove that, for every fragment FO, , the combined complexity of FOMC (or WFOMC) is #P-complete.Comment: To appear at PODS'1
Graphical Models and Symmetries : Loopy Belief Propagation Approaches
Whenever a person or an automated system has to reason in uncertain domains, probability theory is necessary. Probabilistic graphical models allow us to build statistical models that capture complex dependencies between random variables. Inference in these models, however, can easily become intractable. Typical ways to address this scaling issue are inference by approximate message-passing, stochastic gradients, and MapReduce, among others. Exploiting the symmetries of graphical models, however, has not yet been considered for scaling statistical machine learning applications. One instance of graphical models that are inherently symmetric are statistical relational models. These have recently gained attraction within the machine learning and AI communities and combine probability theory with first-order logic, thereby allowing for an efficient representation of structured relational domains. The provided formalisms to compactly represent complex real-world domains enable us to effectively describe large problem instances. Inference within and training of graphical models, however, have not been able to keep pace with the increased representational power. This thesis tackles two major aspects of graphical models and shows that both inference and training can indeed benefit from exploiting symmetries. It first deals with efficient inference exploiting symmetries in graphical models for various query types. We introduce lifted loopy belief propagation (lifted LBP), the first lifted parallel inference approach for relational as well as propositional graphical models. Lifted LBP can effectively speed up marginal inference, but cannot straightforwardly be applied to other types of queries. Thus we also demonstrate efficient lifted algorithms for MAP inference and higher order marginals, as well as the efficient handling of multiple inference tasks. Then we turn to the training of graphical models and introduce the first lifted online training for relational models. Our training procedure and the MapReduce lifting for loopy belief propagation combine lifting with the traditional statistical approaches to scaling, thereby bridging the gap between statistical relational learning and traditional statistical machine learning
Machine Learning for Instance Segmentation
Volumetric Electron Microscopy images can be used for connectomics, the study of brain connectivity at the cellular level.
A prerequisite for this inquiry is the automatic identification of neural cells, which requires machine learning algorithms and in particular efficient image segmentation algorithms.
In this thesis, we develop new algorithms for this task.
In the first part we provide, for the first time in this
field, a method for training a neural network to predict optimal input data for a watershed algorithm.
We demonstrate its superior performance compared to other segmentation methods of its category.
In the second part, we develop an efficient watershed-based algorithm for weighted graph
partitioning, the \emph{Mutex Watershed}, which uses negative edge-weights for the first time.
We show that it is intimately related to the multicut and has a cutting edge performance on a connectomics challenge.
Our algorithm is currently used by the leaders of two connectomics challenges.
Finally, motivated by inpainting neural networks, we create a method to learn the graph weights without any supervision
Open-World Probabilistic Databases: An Abridged Report *
Abstract Large-scale probabilistic knowledge bases are becoming increasingly important in academia and industry alike. They are constantly extended with new data, powered by modern information extraction tools that associate probabilities with database tuples. In this paper, we revisit the semantics underlying such systems. In particular, the closed-world assumption of probabilistic databases, that facts not in the database have probability zero, clearly conflicts with their everyday use. To address this discrepancy, we propose an open-world probabilistic database semantics, which relaxes the probabilities of open facts to default intervals. For this openworld setting, we lift the existing data complexity dichotomy of probabilistic databases, and propose an efficient evaluation algorithm for unions of conjunctive queries. We also show that query evaluation can become harder for non-monotone queries
Gossip Algorithms for Distributed Signal Processing
Gossip algorithms are attractive for in-network processing in sensor networks
because they do not require any specialized routing, there is no bottleneck or
single point of failure, and they are robust to unreliable wireless network
conditions. Recently, there has been a surge of activity in the computer
science, control, signal processing, and information theory communities,
developing faster and more robust gossip algorithms and deriving theoretical
performance guarantees. This article presents an overview of recent work in the
area. We describe convergence rate results, which are related to the number of
transmitted messages and thus the amount of energy consumed in the network for
gossiping. We discuss issues related to gossiping over wireless links,
including the effects of quantization and noise, and we illustrate the use of
gossip algorithms for canonical signal processing tasks including distributed
estimation, source localization, and compression.Comment: Submitted to Proceedings of the IEEE, 29 page
Logical Abstractions for Noisy Variational Quantum Algorithm Simulation
Due to the unreliability and limited capacity of existing quantum computer
prototypes, quantum circuit simulation continues to be a vital tool for
validating next generation quantum computers and for studying variational
quantum algorithms, which are among the leading candidates for useful quantum
computation. Existing quantum circuit simulators do not address the common
traits of variational algorithms, namely: 1) their ability to work with noisy
qubits and operations, 2) their repeated execution of the same circuits but
with different parameters, and 3) the fact that they sample from circuit final
wavefunctions to drive a classical optimization routine. We present a quantum
circuit simulation toolchain based on logical abstractions targeted for
simulating variational algorithms. Our proposed toolchain encodes quantum
amplitudes and noise probabilities in a probabilistic graphical model, and it
compiles the circuits to logical formulas that support efficient repeated
simulation of and sampling from quantum circuits for different parameters.
Compared to state-of-the-art state vector and density matrix quantum circuit
simulators, our simulation approach offers greater performance when sampling
from noisy circuits with at least eight to 20 qubits and with around 12
operations on each qubit, making the approach ideal for simulating near-term
variational quantum algorithms. And for simulating noise-free shallow quantum
circuits with 32 qubits, our simulation approach offers a reduction
in sampling cost versus quantum circuit simulation techniques based on tensor
network contraction.Comment: ASPLOS '21, April 19-23, 2021, Virtual, US
- …