399,720 research outputs found
The thermodynamics of prediction
A system responding to a stochastic driving signal can be interpreted as
computing, by means of its dynamics, an implicit model of the environmental
variables. The system's state retains information about past environmental
fluctuations, and a fraction of this information is predictive of future ones.
The remaining nonpredictive information reflects model complexity that does not
improve predictive power, and thus represents the ineffectiveness of the model.
We expose the fundamental equivalence between this model inefficiency and
thermodynamic inefficiency, measured by dissipation. Our results hold
arbitrarily far from thermodynamic equilibrium and are applicable to a wide
range of systems, including biomolecular machines. They highlight a profound
connection between the effective use of information and efficient thermodynamic
operation: any system constructed to keep memory about its environment and to
operate with maximal energetic efficiency has to be predictive.Comment: 5 pages, 1 figur
Predictability, complexity and learning
We define {\em predictive information} as the mutual
information between the past and the future of a time series. Three
qualitatively different behaviors are found in the limit of large observation
times : can remain finite, grow logarithmically, or grow
as a fractional power law. If the time series allows us to learn a model with a
finite number of parameters, then grows logarithmically with
a coefficient that counts the dimensionality of the model space. In contrast,
power--law growth is associated, for example, with the learning of infinite
parameter (or nonparametric) models such as continuous functions with
smoothness constraints. There are connections between the predictive
information and measures of complexity that have been defined both in learning
theory and in the analysis of physical systems through statistical mechanics
and dynamical systems theory. Further, in the same way that entropy provides
the unique measure of available information consistent with some simple and
plausible conditions, we argue that the divergent part of
provides the unique measure for the complexity of dynamics underlying a time
series. Finally, we discuss how these ideas may be useful in different problems
in physics, statistics, and biology.Comment: 53 pages, 3 figures, 98 references, LaTeX2
Training-free Measures Based on Algorithmic Probability Identify High Nucleosome Occupancy in DNA Sequences
We introduce and study a set of training-free methods of
information-theoretic and algorithmic complexity nature applied to DNA
sequences to identify their potential capabilities to determine nucleosomal
binding sites. We test our measures on well-studied genomic sequences of
different sizes drawn from different sources. The measures reveal the known in
vivo versus in vitro predictive discrepancies and uncover their potential to
pinpoint (high) nucleosome occupancy. We explore different possible signals
within and beyond the nucleosome length and find that complexity indices are
informative of nucleosome occupancy. We compare against the gold standard
(Kaplan model) and find similar and complementary results with the main
difference that our sequence complexity approach. For example, for high
occupancy, complexity-based scores outperform the Kaplan model for predicting
binding representing a significant advancement in predicting the highest
nucleosome occupancy following a training-free approach.Comment: 8 pages main text (4 figures), 12 total with Supplementary (1 figure
Information theory of quantum systems with some hydrogenic applications
The information-theoretic representation of quantum systems, which
complements the familiar energy description of the density-functional and
wave-function-based theories, is here discussed. According to it, the internal
disorder of the quantum-mechanical non-relativistic systems can be quantified
by various single (Fisher information, Shannon entropy) and composite (e.g.
Cramer-Rao, LMC shape and Fisher-Shannon complexity) functionals of the
Schr\"odinger probability density. First, we examine these concepts and its
application to quantum systems with central potentials. Then, we calculate
these measures for hydrogenic systems, emphasizing their predictive power for
various physical phenomena. Finally, some recent open problems are pointed out.Comment: 9 pages, 3 figure
Information theoretic approach to interactive learning
The principles of statistical mechanics and information theory play an
important role in learning and have inspired both theory and the design of
numerous machine learning algorithms. The new aspect in this paper is a focus
on integrating feedback from the learner. A quantitative approach to
interactive learning and adaptive behavior is proposed, integrating model- and
decision-making into one theoretical framework. This paper follows simple
principles by requiring that the observer's world model and action policy
should result in maximal predictive power at minimal complexity. Classes of
optimal action policies and of optimal models are derived from an objective
function that reflects this trade-off between prediction and complexity. The
resulting optimal models then summarize, at different levels of abstraction,
the process's causal organization in the presence of the learner's actions. A
fundamental consequence of the proposed principle is that the learner's optimal
action policies balance exploration and control as an emerging property.
Interestingly, the explorative component is present in the absence of policy
randomness, i.e. in the optimal deterministic behavior. This is a direct result
of requiring maximal predictive power in the presence of feedback.Comment: 6 page
- …