17,824 research outputs found
LISA Source Confusion
The Laser Interferometer Space Antenna (LISA) will detect thousands of
gravitational wave sources. Many of these sources will be overlapping in the
sense that their signals will have a non-zero cross-correlation. Such overlaps
lead to source confusion, which adversely affects how well we can extract
information about the individual sources. Here we study how source confusion
impacts parameter estimation for galactic compact binaries, with emphasis on
the effects of the number of overlaping sources, the time of observation, the
gravitational wave frequencies of the sources, and the degree of the signal
correlations. Our main findings are that the parameter resolution decays
exponentially with the number of overlapping sources, and super-exponentially
with the degree of cross-correlation. We also find that an extended mission
lifetime is key to disentangling the source confusion as the parameter
resolution for overlapping sources improves much faster than the usual square
root of the observation time.Comment: 8 pages, 14 figure
Time's Barbed Arrow: Irreversibility, Crypticity, and Stored Information
We show why the amount of information communicated between the past and
future--the excess entropy--is not in general the amount of information stored
in the present--the statistical complexity. This is a puzzle, and a
long-standing one, since the latter is what is required for optimal prediction,
but the former describes observed behavior. We layout a classification scheme
for dynamical systems and stochastic processes that determines when these two
quantities are the same or different. We do this by developing closed-form
expressions for the excess entropy in terms of optimal causal predictors and
retrodictors--the epsilon-machines of computational mechanics. A process's
causal irreversibility and crypticity are key determining properties.Comment: 4 pages, 2 figure
The Minimum Description Length Principle and Model Selection in Spectropolarimetry
It is shown that the two-part Minimum Description Length Principle can be
used to discriminate among different models that can explain a given observed
dataset. The description length is chosen to be the sum of the lengths of the
message needed to encode the model plus the message needed to encode the data
when the model is applied to the dataset. It is verified that the proposed
principle can efficiently distinguish the model that correctly fits the
observations while avoiding over-fitting. The capabilities of this criterion
are shown in two simple problems for the analysis of observed
spectropolarimetric signals. The first is the de-noising of observations with
the aid of the PCA technique. The second is the selection of the optimal number
of parameters in LTE inversions. We propose this criterion as a quantitative
approach for distinguising the most plausible model among a set of proposed
models. This quantity is very easy to implement as an additional output on the
existing inversion codes.Comment: Accepted for publication in the Astrophysical Journa
Fluctuation Theorem with Information Exchange: Role of Correlations in Stochastic Thermodynamics
We establish the fluctuation theorem in the presence of information exchange
between a nonequilibrium system and other degrees of freedom such as an
observer and a feedback controller, where the amount of information exchange is
added to the entropy production. The resulting generalized second law sets the
fundamental limit of energy dissipation and energy cost during the information
exchange. Our results apply not only to feedback-controlled processes but also
to a much broader class of information exchanges, and provides a unified
framework of nonequilibrium thermodynamics of measurement and feedback control.Comment: To appear in PR
Entropy exchange and entanglement in the Jaynes-Cummings model
The Jaynes-Cummings model is the simplest fully quantum model that describes
the interaction between light and matter. We extend a previous analysis by
Phoenix and Knight (S. J. D. Phoenix, P. L. Knight, Annals of Physics 186,
381). of the JCM by considering mixed states of both the light and matter. We
present examples of qualitatively different entropic correlations. In
particular, we explore the regime of entropy exchange between light and matter,
i.e. where the rate of change of the two are anti-correlated. This behavior
contrasts with the case of pure light-matter states in which the rate of change
of the two entropies are positively correlated and in fact identical. We give
an analytical derivation of the anti-correlation phenomenon and discuss the
regime of its validity. Finally, we show a strong correlation between the
region of the Bloch sphere characterized by entropy exchange and that
characterized by minimal entanglement as measured by the negative eigenvalues
of the partially transposed density matrix.Comment: 8 pages, 5 figure
Measuring the effective complexity of cosmological models
We introduce a statistical measure of the effective model complexity, called
the Bayesian complexity. We demonstrate that the Bayesian complexity can be
used to assess how many effective parameters a set of data can support and that
it is a useful complement to the model likelihood (the evidence) in model
selection questions. We apply this approach to recent measurements of cosmic
microwave background anisotropies combined with the Hubble Space Telescope
measurement of the Hubble parameter. Using mildly non-informative priors, we
show how the 3-year WMAP data improves on the first-year data by being able to
measure both the spectral index and the reionization epoch at the same time. We
also find that a non-zero curvature is strongly disfavored. We conclude that
although current data could constrain at least seven effective parameters, only
six of them are required in a scheme based on the Lambda-CDM concordance
cosmology.Comment: 9 pages, 4 figures, revised version accepted for publication in PRD,
updated with WMAP3 result
Generalized Hurst exponent and multifractal function of original and translated texts mapped into frequency and length time series
A nonlinear dynamics approach can be used in order to quantify complexity in
written texts. As a first step, a one-dimensional system is examined : two
written texts by one author (Lewis Carroll) are considered, together with one
translation, into an artificial language, i.e. Esperanto are mapped into time
series. Their corresponding shuffled versions are used for obtaining a "base
line". Two different one-dimensional time series are used here: (i) one based
on word lengths (LTS), (ii) the other on word frequencies (FTS). It is shown
that the generalized Hurst exponent and the derived curves
of the original and translated texts show marked differences. The original
"texts" are far from giving a parabolic function, - in contrast to
the shuffled texts. Moreover, the Esperanto text has more extreme values. This
suggests cascade model-like, with multiscale time asymmetric features as
finally written texts. A discussion of the difference and complementarity of
mapping into a LTS or FTS is presented. The FTS curves are more
opened than the LTS onesComment: preprint for PRE; 2 columns; 10 pages; 6 (multifigures); 3 Tables; 70
reference
Statistical mechanical aspects of joint source-channel coding
An MN-Gallager Code over Galois fields, , based on the Dynamical Block
Posterior probabilities (DBP) for messages with a given set of autocorrelations
is presented with the following main results: (a) for a binary symmetric
channel the threshold, , is extrapolated for infinite messages using the
scaling relation for the median convergence time, ;
(b) a degradation in the threshold is observed as the correlations are
enhanced; (c) for a given set of autocorrelations the performance is enhanced
as is increased; (d) the efficiency of the DBP joint source-channel coding
is slightly better than the standard gzip compression method; (e) for a given
entropy, the performance of the DBP algorithm is a function of the decay of the
correlation function over large distances.Comment: 6 page
Maximum Entropy and Bayesian Data Analysis: Entropic Priors
The problem of assigning probability distributions which objectively reflect
the prior information available about experiments is one of the major stumbling
blocks in the use of Bayesian methods of data analysis. In this paper the
method of Maximum (relative) Entropy (ME) is used to translate the information
contained in the known form of the likelihood into a prior distribution for
Bayesian inference. The argument is inspired and guided by intuition gained
from the successful use of ME methods in statistical mechanics. For experiments
that cannot be repeated the resulting "entropic prior" is formally identical
with the Einstein fluctuation formula. For repeatable experiments, however, the
expected value of the entropy of the likelihood turns out to be relevant
information that must be included in the analysis. The important case of a
Gaussian likelihood is treated in detail.Comment: 23 pages, 2 figure
- …