4,952 research outputs found
Neural Probabilistic Logic Programming in Discrete-Continuous Domains
Neural-symbolic AI (NeSy) allows neural networks to exploit symbolic
background knowledge in the form of logic. It has been shown to aid learning in
the limited data regime and to facilitate inference on out-of-distribution
data. Probabilistic NeSy focuses on integrating neural networks with both logic
and probability theory, which additionally allows learning under uncertainty. A
major limitation of current probabilistic NeSy systems, such as DeepProbLog, is
their restriction to finite probability distributions, i.e., discrete random
variables. In contrast, deep probabilistic programming (DPP) excels in
modelling and optimising continuous probability distributions. Hence, we
introduce DeepSeaProbLog, a neural probabilistic logic programming language
that incorporates DPP techniques into NeSy. Doing so results in the support of
inference and learning of both discrete and continuous probability
distributions under logical constraints. Our main contributions are 1) the
semantics of DeepSeaProbLog and its corresponding inference algorithm, 2) a
proven asymptotically unbiased learning algorithm, and 3) a series of
experiments that illustrate the versatility of our approach.Comment: 27 pages, 9 figure
Symbolic Logic meets Machine Learning: A Brief Survey in Infinite Domains
The tension between deduction and induction is perhaps the most fundamental
issue in areas such as philosophy, cognition and artificial intelligence (AI).
The deduction camp concerns itself with questions about the expressiveness of
formal languages for capturing knowledge about the world, together with proof
systems for reasoning from such knowledge bases. The learning camp attempts to
generalize from examples about partial descriptions about the world. In AI,
historically, these camps have loosely divided the development of the field,
but advances in cross-over areas such as statistical relational learning,
neuro-symbolic systems, and high-level control have illustrated that the
dichotomy is not very constructive, and perhaps even ill-formed. In this
article, we survey work that provides further evidence for the connections
between logic and learning. Our narrative is structured in terms of three
strands: logic versus learning, machine learning for logic, and logic for
machine learning, but naturally, there is considerable overlap. We place an
emphasis on the following "sore" point: there is a common misconception that
logic is for discrete properties, whereas probability theory and machine
learning, more generally, is for continuous properties. We report on results
that challenge this view on the limitations of logic, and expose the role that
logic can play for learning in infinite domains
Numeric Input Relations for Relational Learning with Applications to Community Structure Analysis
Most work in the area of statistical relational learning (SRL) is focussed on
discrete data, even though a few approaches for hybrid SRL models have been
proposed that combine numerical and discrete variables. In this paper we
distinguish numerical random variables for which a probability distribution is
defined by the model from numerical input variables that are only used for
conditioning the distribution of discrete response variables. We show how
numerical input relations can very easily be used in the Relational Bayesian
Network framework, and that existing inference and learning methods need only
minor adjustments to be applied in this generalized setting. The resulting
framework provides natural relational extensions of classical probabilistic
models for categorical data. We demonstrate the usefulness of RBN models with
numeric input relations by several examples.
In particular, we use the augmented RBN framework to define probabilistic
models for multi-relational (social) networks in which the probability of a
link between two nodes depends on numeric latent feature vectors associated
with the nodes. A generic learning procedure can be used to obtain a
maximum-likelihood fit of model parameters and latent feature values for a
variety of models that can be expressed in the high-level RBN representation.
Specifically, we propose a model that allows us to interpret learned latent
feature values as community centrality degrees by which we can identify nodes
that are central for one community, that are hubs between communities, or that
are isolated nodes. In a multi-relational setting, the model also provides a
characterization of how different relations are associated with each community
- …