3,201 research outputs found
Exact mean field inference in asymmetric kinetic Ising systems
We develop an elementary mean field approach for fully asymmetric kinetic
Ising models, which can be applied to a single instance of the problem. In the
case of the asymmetric SK model this method gives the exact values of the local
magnetizations and the exact relation between equal-time and time-delayed
correlations. It can also be used to solve efficiently the inverse problem,
i.e. determine the couplings and local fields from a set of patterns, also in
cases where the fields and couplings are time-dependent. This approach
generalizes some recent attempts to solve this dynamical inference problem,
which were valid in the limit of weak coupling. It provides the exact solution
to the problem also in strongly coupled problems. This mean field inference can
also be used as an efficient approximate method to infer the couplings and
fields in problems which are not infinite range, for instance in diluted
asymmetric spin glasses.Comment: 10 pages, 7 figure
Sentient Networks
In this paper we consider the question whether a distributed network of
sensors and data processors can form "perceptions" based on the sensory data.
Because sensory data can have exponentially many explanations, the use of a
central data processor to analyze the outputs from a large ensemble of sensors
will in general introduce unacceptable latencies for responding to dangerous
situations. A better idea is to use a distributed "Helmholtz machine"
architecture in which the collective state of the network as a whole provides
an explanation for the sensory data.Comment: PostScript, 14 page
Dreaming neural networks: forgetting spurious memories and reinforcing pure ones
The standard Hopfield model for associative neural networks accounts for
biological Hebbian learning and acts as the harmonic oscillator for pattern
recognition, however its maximal storage capacity is , far
from the theoretical bound for symmetric networks, i.e. . Inspired
by sleeping and dreaming mechanisms in mammal brains, we propose an extension
of this model displaying the standard on-line (awake) learning mechanism (that
allows the storage of external information in terms of patterns) and an
off-line (sleep) unlearningconsolidating mechanism (that allows
spurious-pattern removal and pure-pattern reinforcement): this obtained daily
prescription is able to saturate the theoretical bound , remaining
also extremely robust against thermal noise. Both neural and synaptic features
are analyzed both analytically and numerically. In particular, beyond obtaining
a phase diagram for neural dynamics, we focus on synaptic plasticity and we
give explicit prescriptions on the temporal evolution of the synaptic matrix.
We analytically prove that our algorithm makes the Hebbian kernel converge with
high probability to the projection matrix built over the pure stored patterns.
Furthermore, we obtain a sharp and explicit estimate for the "sleep rate" in
order to ensure such a convergence. Finally, we run extensive numerical
simulations (mainly Monte Carlo sampling) to check the approximations
underlying the analytical investigations (e.g., we developed the whole theory
at the so called replica-symmetric level, as standard in the
Amit-Gutfreund-Sompolinsky reference framework) and possible finite-size
effects, finding overall full agreement with the theory.Comment: 31 pages, 12 figure
Stochastic Synapses Enable Efficient Brain-Inspired Learning Machines
Recent studies have shown that synaptic unreliability is a robust and
sufficient mechanism for inducing the stochasticity observed in cortex. Here,
we introduce Synaptic Sampling Machines, a class of neural network models that
uses synaptic stochasticity as a means to Monte Carlo sampling and unsupervised
learning. Similar to the original formulation of Boltzmann machines, these
models can be viewed as a stochastic counterpart of Hopfield networks, but
where stochasticity is induced by a random mask over the connections. Synaptic
stochasticity plays the dual role of an efficient mechanism for sampling, and a
regularizer during learning akin to DropConnect. A local synaptic plasticity
rule implementing an event-driven form of contrastive divergence enables the
learning of generative models in an on-line fashion. Synaptic sampling machines
perform equally well using discrete-timed artificial units (as in Hopfield
networks) or continuous-timed leaky integrate & fire neurons. The learned
representations are remarkably sparse and robust to reductions in bit precision
and synapse pruning: removal of more than 75% of the weakest connections
followed by cursory re-learning causes a negligible performance loss on
benchmark classification tasks. The spiking neuron-based synaptic sampling
machines outperform existing spike-based unsupervised learners, while
potentially offering substantial advantages in terms of power and complexity,
and are thus promising models for on-line learning in brain-inspired hardware
Minimum Energy Information Fusion in Sensor Networks
In this paper we consider how to organize the sharing of information in a
distributed network of sensors and data processors so as to provide
explanations for sensor readings with minimal expenditure of energy. We point
out that the Minimum Description Length principle provides an approach to
information fusion that is more naturally suited to energy minimization than
traditional Bayesian approaches. In addition we show that for networks
consisting of a large number of identical sensors Kohonen self-organization
provides an exact solution to the problem of combining the sensor outputs into
minimal description length explanations.Comment: postscript, 8 pages. Paper 65 in Proceedings of The 2nd International
Conference on Information Fusio
- …