46,493 research outputs found
Specific and complete local integration of patterns in Bayesian networks
© 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). The Version of Record, Daniel Polani, et al, Entropy, Vol. 19 (5), Article number 230, published 18 May 2017, is available online at doi:10.3390/e19050230.We present a first formal analysis of specific and complete local integration. Complete local integration was previously proposed as a criterion for detecting entities or wholes in distributed dynamical systems. Such entities in turn were conceived to form the basis of a theory of emergence of agents within dynamical systems. Here, we give a more thorough account of the underlying formal measures. The main contribution is the disintegration theorem which reveals a special role of completely locally integrated patterns (what we call ι-entities) within the trajectories they occur in. Apart from proving this theorem we introduce the disintegration hierarchy and its refinement-free version as a way to structure the patterns in a trajectory. Furthermore we construct the least upper bound and provide a candidate for the greatest lower bound of specific local integration. Finally, we calculate the ι-entities in small example systems as a first sanity check and find that ι-entities largely fulfil simple expectations.Peer reviewedFinal Published versio
Towards a Multi-Subject Analysis of Neural Connectivity
Directed acyclic graphs (DAGs) and associated probability models are widely
used to model neural connectivity and communication channels. In many
experiments, data are collected from multiple subjects whose connectivities may
differ but are likely to share many features. In such circumstances it is
natural to leverage similarity between subjects to improve statistical
efficiency. The first exact algorithm for estimation of multiple related DAGs
was recently proposed by Oates et al. 2014; in this letter we present examples
and discuss implications of the methodology as applied to the analysis of fMRI
data from a multi-subject experiment. Elicitation of tuning parameters requires
care and we illustrate how this may proceed retrospectively based on technical
replicate data. In addition to joint learning of subject-specific connectivity,
we allow for heterogeneous collections of subjects and simultaneously estimate
relationships between the subjects themselves. This letter aims to highlight
the potential for exact estimation in the multi-subject setting.Comment: to appear in Neural Computation 27:1-2
Network Plasticity as Bayesian Inference
General results from statistical learning theory suggest to understand not
only brain computations, but also brain plasticity as probabilistic inference.
But a model for that has been missing. We propose that inherently stochastic
features of synaptic plasticity and spine motility enable cortical networks of
neurons to carry out probabilistic inference by sampling from a posterior
distribution of network configurations. This model provides a viable
alternative to existing models that propose convergence of parameters to
maximum likelihood values. It explains how priors on weight distributions and
connection probabilities can be merged optimally with learned experience, how
cortical networks can generalize learned information so well to novel
experiences, and how they can compensate continuously for unforeseen
disturbances of the network. The resulting new theory of network plasticity
explains from a functional perspective a number of experimental data on
stochastic aspects of synaptic plasticity that previously appeared to be quite
puzzling.Comment: 33 pages, 5 figures, the supplement is available on the author's web
page http://www.igi.tugraz.at/kappe
- …