691 research outputs found
Two remarks on generalized entropy power inequalities
This note contributes to the understanding of generalized entropy power
inequalities. Our main goal is to construct a counter-example regarding
monotonicity and entropy comparison of weighted sums of independent identically
distributed log-concave random variables. We also present a complex analogue of
a recent dependent entropy power inequality of Hao and Jog, and give a very
simple proof.Comment: arXiv:1811.00345 is split into 2 papers, with this being on
Information-Theoretic Analysis of Serial Dependence and Cointegration.
This paper is devoted to presenting wider characterizations of memory and cointegration in time series, in terms of information-theoretic statistics such as the entropy and the mutual information between pairs of variables. We suggest a nonparametric and nonlinear methodology for data analysis and for testing the hypotheses of long memory and the existence of a cointegrating relationship in a nonlinear context. This new framework represents a natural extension of the linear-memory concepts based on correlations. Finally, we show that our testing devices seem promising for exploratory analysis with nonlinearly cointegrated time series.
The large deviation approach to statistical mechanics
The theory of large deviations is concerned with the exponential decay of
probabilities of large fluctuations in random systems. These probabilities are
important in many fields of study, including statistics, finance, and
engineering, as they often yield valuable information about the large
fluctuations of a random system around its most probable state or trajectory.
In the context of equilibrium statistical mechanics, the theory of large
deviations provides exponential-order estimates of probabilities that refine
and generalize Einstein's theory of fluctuations. This review explores this and
other connections between large deviation theory and statistical mechanics, in
an effort to show that the mathematical language of statistical mechanics is
the language of large deviation theory. The first part of the review presents
the basics of large deviation theory, and works out many of its classical
applications related to sums of random variables and Markov processes. The
second part goes through many problems and results of statistical mechanics,
and shows how these can be formulated and derived within the context of large
deviation theory. The problems and results treated cover a wide range of
physical systems, including equilibrium many-particle systems, noise-perturbed
dynamics, nonequilibrium systems, as well as multifractals, disordered systems,
and chaotic systems. This review also covers many fundamental aspects of
statistical mechanics, such as the derivation of variational principles
characterizing equilibrium and nonequilibrium states, the breaking of the
Legendre transform for nonconcave entropies, and the characterization of
nonequilibrium fluctuations through fluctuation relations.Comment: v1: 89 pages, 18 figures, pdflatex. v2: 95 pages, 20 figures, text,
figures and appendices added, many references cut, close to published versio
Information-Theoretic Analysis of Serial Dependence and Cointegration
This paper is devoted to presenting wider characterizations of memory and cointegration in time series, in terms of information-theoretic statistics such as the entropy and the mutual information between pairs of variables. We suggest a nonparametric and nonlinear methodology for data analysis and for testing the hypotheses of long memory and the existence of a cointegrating relationship in a nonlinear context. This new framework represents a natural extension of the linear-memory concepts based on correlations. Finally, we show that our testing devices seem promising for exploratory analysis with nonlinearly cointegrated time series.Publicad
- …