3,220 research outputs found
The Computational Structure of Spike Trains
Neurons perform computations, and convey the results of those computations
through the statistical structure of their output spike trains. Here we present
a practical method, grounded in the information-theoretic analysis of
prediction, for inferring a minimal representation of that structure and for
characterizing its complexity. Starting from spike trains, our approach finds
their causal state models (CSMs), the minimal hidden Markov models or
stochastic automata capable of generating statistically identical time series.
We then use these CSMs to objectively quantify both the generalizable structure
and the idiosyncratic randomness of the spike train. Specifically, we show that
the expected algorithmic information content (the information needed to
describe the spike train exactly) can be split into three parts describing (1)
the time-invariant structure (complexity) of the minimal spike-generating
process, which describes the spike train statistically; (2) the randomness
(internal entropy rate) of the minimal spike-generating process; and (3) a
residual pure noise term not described by the minimal spike-generating process.
We use CSMs to approximate each of these quantities. The CSMs are inferred
nonparametrically from the data, making only mild regularity assumptions, via
the causal state splitting reconstruction algorithm. The methods presented here
complement more traditional spike train analyses by describing not only spiking
probability and spike train entropy, but also the complexity of a spike train's
structure. We demonstrate our approach using both simulated spike trains and
experimental data recorded in rat barrel cortex during vibrissa stimulation.Comment: Somewhat different format from journal version but same conten
Signatures of Infinity: Nonergodicity and Resource Scaling in Prediction, Complexity, and Learning
We introduce a simple analysis of the structural complexity of
infinite-memory processes built from random samples of stationary, ergodic
finite-memory component processes. Such processes are familiar from the well
known multi-arm Bandit problem. We contrast our analysis with
computation-theoretic and statistical inference approaches to understanding
their complexity. The result is an alternative view of the relationship between
predictability, complexity, and learning that highlights the distinct ways in
which informational and correlational divergences arise in complex ergodic and
nonergodic processes. We draw out consequences for the resource divergences
that delineate the structural hierarchy of ergodic processes and for processes
that are themselves hierarchical.Comment: 8 pages, 1 figure; http://csc.ucdavis.edu/~cmg/compmech/pubs/soi.pd
Quantum walks: a comprehensive review
Quantum walks, the quantum mechanical counterpart of classical random walks,
is an advanced tool for building quantum algorithms that has been recently
shown to constitute a universal model of quantum computation. Quantum walks is
now a solid field of research of quantum computation full of exciting open
problems for physicists, computer scientists, mathematicians and engineers.
In this paper we review theoretical advances on the foundations of both
discrete- and continuous-time quantum walks, together with the role that
randomness plays in quantum walks, the connections between the mathematical
models of coined discrete quantum walks and continuous quantum walks, the
quantumness of quantum walks, a summary of papers published on discrete quantum
walks and entanglement as well as a succinct review of experimental proposals
and realizations of discrete-time quantum walks. Furthermore, we have reviewed
several algorithms based on both discrete- and continuous-time quantum walks as
well as a most important result: the computational universality of both
continuous- and discrete- time quantum walks.Comment: Paper accepted for publication in Quantum Information Processing
Journa
- …