16,821 research outputs found
Logarithmic Time Parallel Bayesian Inference
I present a parallel algorithm for exact probabilistic inference in Bayesian
networks. For polytree networks with n variables, the worst-case time
complexity is O(log n) on a CREW PRAM (concurrent-read, exclusive-write
parallel random-access machine) with n processors, for any constant number of
evidence variables. For arbitrary networks, the time complexity is O(r^{3w}*log
n) for n processors, or O(w*log n) for r^{3w}*n processors, where r is the
maximum range of any variable, and w is the induced width (the maximum clique
size), after moralizing and triangulating the network.Comment: Appears in Proceedings of the Fourteenth Conference on Uncertainty in
Artificial Intelligence (UAI1998
Magnetic Tunnel Junction Mimics Stochastic Cortical Spiking Neurons
Brain-inspired computing architectures attempt to mimic the computations
performed in the neurons and the synapses in the human brain in order to
achieve its efficiency in learning and cognitive tasks. In this work, we
demonstrate the mapping of the probabilistic spiking nature of pyramidal
neurons in the cortex to the stochastic switching behavior of a Magnetic Tunnel
Junction in presence of thermal noise. We present results to illustrate the
efficiency of neuromorphic systems based on such probabilistic neurons for
pattern recognition tasks in presence of lateral inhibition and homeostasis.
Such stochastic MTJ neurons can also potentially provide a direct mapping to
the probabilistic computing elements in Belief Networks for performing
regenerative tasks.Comment: The article will appear in Scientific Report
Bayesian Network Constraint-Based Structure Learning Algorithms: Parallel and Optimised Implementations in the bnlearn R Package
It is well known in the literature that the problem of learning the structure
of Bayesian networks is very hard to tackle: its computational complexity is
super-exponential in the number of nodes in the worst case and polynomial in
most real-world scenarios.
Efficient implementations of score-based structure learning benefit from past
and current research in optimisation theory, which can be adapted to the task
by using the network score as the objective function to maximise. This is not
true for approaches based on conditional independence tests, called
constraint-based learning algorithms. The only optimisation in widespread use,
backtracking, leverages the symmetries implied by the definitions of
neighbourhood and Markov blanket.
In this paper we illustrate how backtracking is implemented in recent
versions of the bnlearn R package, and how it degrades the stability of
Bayesian network structure learning for little gain in terms of speed. As an
alternative, we describe a software architecture and framework that can be used
to parallelise constraint-based structure learning algorithms (also implemented
in bnlearn) and we demonstrate its performance using four reference networks
and two real-world data sets from genetics and systems biology. We show that on
modern multi-core or multiprocessor hardware parallel implementations are
preferable over backtracking, which was developed when single-processor
machines were the norm.Comment: 20 pages, 4 figure
Cognitive computational neuroscience
To learn how cognition is implemented in the brain, we must build
computational models that can perform cognitive tasks, and test such models
with brain and behavioral experiments. Cognitive science has developed
computational models of human cognition, decomposing task performance into
computational components. However, its algorithms still fall short of human
intelligence and are not grounded in neurobiology. Computational neuroscience
has investigated how interacting neurons can implement component functions of
brain computation. However, it has yet to explain how those components interact
to explain human cognition and behavior. Modern technologies enable us to
measure and manipulate brain activity in unprecedentedly rich ways in animals
and humans. However, experiments will yield theoretical insight only when
employed to test brain-computational models. It is time to assemble the pieces
of the puzzle of brain computation. Here we review recent work in the
intersection of cognitive science, computational neuroscience, and artificial
intelligence. Computational models that mimic brain information processing
during perceptual, cognitive, and control tasks are beginning to be developed
and tested with brain and behavioral data.Comment: 31 pages, 4 figure
An Application of Uncertain Reasoning to Requirements Engineering
This paper examines the use of Bayesian Networks to tackle one of the tougher
problems in requirements engineering, translating user requirements into system
requirements. The approach taken is to model domain knowledge as Bayesian
Network fragments that are glued together to form a complete view of the domain
specific system requirements. User requirements are introduced as evidence and
the propagation of belief is used to determine what are the appropriate system
requirements as indicated by user requirements. This concept has been
demonstrated in the development of a system specification and the results are
presented here.Comment: Appears in Proceedings of the Fifteenth Conference on Uncertainty in
Artificial Intelligence (UAI1999
ZhuSuan: A Library for Bayesian Deep Learning
In this paper we introduce ZhuSuan, a python probabilistic programming
library for Bayesian deep learning, which conjoins the complimentary advantages
of Bayesian methods and deep learning. ZhuSuan is built upon Tensorflow. Unlike
existing deep learning libraries, which are mainly designed for deterministic
neural networks and supervised tasks, ZhuSuan is featured for its deep root
into Bayesian inference, thus supporting various kinds of probabilistic models,
including both the traditional hierarchical Bayesian models and recent deep
generative models. We use running examples to illustrate the probabilistic
programming on ZhuSuan, including Bayesian logistic regression, variational
auto-encoders, deep sigmoid belief networks and Bayesian recurrent neural
networks.Comment: The GitHub page is at https://github.com/thu-ml/zhusua
Active Neural Localization
Localization is the problem of estimating the location of an autonomous agent
from an observation and a map of the environment. Traditional methods of
localization, which filter the belief based on the observations, are
sub-optimal in the number of steps required, as they do not decide the actions
taken by the agent. We propose "Active Neural Localizer", a fully
differentiable neural network that learns to localize accurately and
efficiently. The proposed model incorporates ideas of traditional
filtering-based localization methods, by using a structured belief of the state
with multiplicative interactions to propagate belief, and combines it with a
policy model to localize accurately while minimizing the number of steps
required for localization. Active Neural Localizer is trained end-to-end with
reinforcement learning. We use a variety of simulation environments for our
experiments which include random 2D mazes, random mazes in the Doom game engine
and a photo-realistic environment in the Unreal game engine. The results on the
2D environments show the effectiveness of the learned policy in an idealistic
setting while results on the 3D environments demonstrate the model's capability
of learning the policy and perceptual model jointly from raw-pixel based RGB
observations. We also show that a model trained on random textures in the Doom
environment generalizes well to a photo-realistic office space environment in
the Unreal engine.Comment: Under Review at ICLR-18, 15 pages, 7 figure
GALGO: A Genetic ALGOrithm Decision Support Tool for Complex Uncertain Systems Modeled with Bayesian Belief Networks
Bayesian belief networks can be used to represent and to reason about complex
systems with uncertain, incomplete and conflicting information. Belief networks
are graphs encoding and quantifying probabilistic dependence and conditional
independence among variables. One type of reasoning of interest in diagnosis is
called abductive inference (determination of the global most probable system
description given the values of any partial subset of variables). In some
cases, abductive inference can be performed with exact algorithms using
distributed network computations but it is an NP-hard problem and complexity
increases drastically with the presence of undirected cycles, number of
discrete states per variable, and number of variables in the network. This
paper describes an approximate method based on genetic algorithms to perform
abductive inference in large, multiply connected networks for which complexity
is a concern when using most exact methods and for which systematic search
methods are not feasible. The theoretical adequacy of the method is discussed
and preliminary experimental results are presented.Comment: Appears in Proceedings of the Ninth Conference on Uncertainty in
Artificial Intelligence (UAI1993
Managing engineering systems with large state and action spaces through deep reinforcement learning
Decision-making for engineering systems can be efficiently formulated as a
Markov Decision Process (MDP) or a Partially Observable MDP (POMDP). Typical
MDP and POMDP solution procedures utilize offline knowledge about the
environment and provide detailed policies for relatively small systems with
tractable state and action spaces. However, in large multi-component systems
the sizes of these spaces easily explode, as system states and actions scale
exponentially with the number of components, whereas environment dynamics are
difficult to be described in explicit forms for the entire system and may only
be accessible through numerical simulators. In this work, to address these
issues, an integrated Deep Reinforcement Learning (DRL) framework is
introduced. The Deep Centralized Multi-agent Actor Critic (DCMAC) is developed,
an off-policy actor-critic DRL approach, providing efficient life-cycle
policies for large multi-component systems operating in high-dimensional
spaces. Apart from deep function approximations that parametrize large state
spaces, DCMAC also adopts a factorized representation of the system actions,
being able to designate individualized component- and subsystem-level
decisions, while maintaining a centralized value function for the entire
system. DCMAC compares well against Deep Q-Network (DQN) solutions and exact
policies, where applicable, and outperforms optimized baselines that are based
on time-based, condition-based and periodic policies
p-Bits for Probabilistic Spin Logic
We introduce the concept of a probabilistic or p-bit, intermediate between
the standard bits of digital electronics and the emerging q-bits of quantum
computing. We show that low barrier magnets or LBM's provide a natural physical
representation for p-bits and can be built either from perpendicular magnets
(PMA) designed to be close to the in-plane transition or from circular in-plane
magnets (IMA). Magnetic tunnel junctions (MTJ) built using LBM's as free layers
can be combined with standard NMOS transistors to provide three-terminal
building blocks for large scale probabilistic circuits that can be designed to
perform useful functions. Interestingly, this three-terminal unit looks just
like the 1T/MTJ device used in embedded MRAM technology, with only one
difference: the use of an LBM for the MTJ free layer. We hope that the concept
of p-bits and p-circuits will help open up new application spaces for this
emerging technology. However, a p-bit need not involve an MTJ, any fluctuating
resistor could be combined with a transistor to implement it, while completely
digital implementations using conventional CMOS technology are also possible.
The p-bit also provides a conceptual bridge between two active but disjoint
fields of research, namely stochastic machine learning and quantum computing.
First, there are the applications that are based on the similarity of a p-bit
to the binary stochastic neuron (BSN), a well-known concept in machine
learning. Three-terminal p-bits could provide an efficient hardware accelerator
for the BSN. Second, there are the applications that are based on the p-bit
being like a poor man's q-bit. Initial demonstrations based on full SPICE
simulations show that several optimization problems including quantum annealing
are amenable to p-bit implementations which can be scaled up at room
temperature using existing technology
- …