1,536 research outputs found
Information Entropy in Cosmology
The effective evolution of an inhomogeneous cosmological model may be
described in terms of spatially averaged variables. We point out that in this
context, quite naturally, a measure arises which is identical to a fluid model
of the `Kullback-Leibler Relative Information Entropy', expressing the
distinguishability of the local inhomogeneous mass density field from its
spatial average on arbitrary compact domains. We discuss the time-evolution of
`effective information' and explore some implications. We conjecture that the
information content of the Universe -- measured by Relative Information Entropy
of a cosmological model containing dust matter -- is increasing.Comment: LateX, PRLstyle, 4 pages; to appear in PR
How `hot' are mixed quantum states?
Given a mixed quantum state of a qudit, we consider any observable
as a kind of `thermometer' in the following sense. Given a source which emits
pure states with these or those distributions, we select such distributions
that the appropriate average value of the observable is equal to the
average Tr of in the stare . Among those distributions we find
the most typical one, namely, having the highest differential entropy. We call
this distribution conditional Gibbs ensemble as it turns out to be a Gibbs
distribution characterized by a temperature-like parameter . The
expressions establishing the liaisons between the density operator and
its temperature parameter are provided. Within this approach, the
uniform mixed state has the highest `temperature', which tends to zero as the
state in question approaches to a pure state.Comment: Contribution to Quantum 2006: III workshop ad memoriam of Carlo
Novero: Advances in Foundations of Quantum Mechanics and Quantum Information
with atoms and photons. 2-5 May 2006 - Turin, Ital
Universality of optimal measurements
We present optimal and minimal measurements on identical copies of an unknown
state of a qubit when the quality of measuring strategies is quantified with
the gain of information (Kullback of probability distributions). We also show
that the maximal gain of information occurs, among isotropic priors, when the
state is known to be pure. Universality of optimal measurements follows from
our results: using the fidelity or the gain of information, two different
figures of merits, leads to exactly the same conclusions. We finally
investigate the optimal capacity of copies of an unknown state as a quantum
channel of information.Comment: Revtex, 5 pages, no figure
Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence
Incremental learning (IL) has received a lot of attention recently, however,
the literature lacks a precise problem definition, proper evaluation settings,
and metrics tailored specifically for the IL problem. One of the main
objectives of this work is to fill these gaps so as to provide a common ground
for better understanding of IL. The main challenge for an IL algorithm is to
update the classifier whilst preserving existing knowledge. We observe that, in
addition to forgetting, a known issue while preserving knowledge, IL also
suffers from a problem we call intransigence, inability of a model to update
its knowledge. We introduce two metrics to quantify forgetting and
intransigence that allow us to understand, analyse, and gain better insights
into the behaviour of IL algorithms. We present RWalk, a generalization of
EWC++ (our efficient version of EWC [Kirkpatrick2016EWC]) and Path Integral
[Zenke2017Continual] with a theoretically grounded KL-divergence based
perspective. We provide a thorough analysis of various IL algorithms on MNIST
and CIFAR-100 datasets. In these experiments, RWalk obtains superior results in
terms of accuracy, and also provides a better trade-off between forgetting and
intransigence
Comparing compact binary parameter distributions I: Methods
Being able to measure each merger's sky location, distance, component masses,
and conceivably spins, ground-based gravitational-wave detectors will provide a
extensive and detailed sample of coalescing compact binaries (CCBs) in the
local and, with third-generation detectors, distant universe. These
measurements will distinguish between competing progenitor formation models. In
this paper we develop practical tools to characterize the amount of
experimentally accessible information available, to distinguish between two a
priori progenitor models. Using a simple time-independent model, we demonstrate
the information content scales strongly with the number of observations. The
exact scaling depends on how significantly mass distributions change between
similar models. We develop phenomenological diagnostics to estimate how many
models can be distinguished, using first-generation and future instruments.
Finally, we emphasize that multi-observable distributions can be fully
exploited only with very precisely calibrated detectors, search pipelines,
parameter estimation, and Bayesian model inference
Quantifying the complexity of random Boolean networks
We study two measures of the complexity of heterogeneous extended systems,
taking random Boolean networks as prototypical cases. A measure defined by
Shalizi et al. for cellular automata, based on a criterion for optimal
statistical prediction [Shalizi et al., Phys. Rev. Lett. 93, 118701 (2004)],
does not distinguish between the spatial inhomogeneity of the ordered phase and
the dynamical inhomogeneity of the disordered phase. A modification in which
complexities of individual nodes are calculated yields vanishing complexity
values for networks in the ordered and critical regimes and for highly
disordered networks, peaking somewhere in the disordered regime. Individual
nodes with high complexity are the ones that pass the most information from the
past to the future, a quantity that depends in a nontrivial way on both the
Boolean function of a given node and its location within the network.Comment: 8 pages, 4 figure
Quantum estimation via minimum Kullback entropy principle
We address quantum estimation in situations where one has at disposal data
from the measurement of an incomplete set of observables and some a priori
information on the state itself. By expressing the a priori information in
terms of a bias toward a given state the problem may be faced by minimizing the
quantum relative entropy (Kullback entropy) with the constraint of reproducing
the data. We exploit the resulting minimum Kullback entropy principle for the
estimation of a quantum state from the measurement of a single observable,
either from the sole mean value or from the complete probability distribution,
and apply it as a tool for the estimation of weak Hamiltonian processes. Qubit
and harmonic oscillator systems are analyzed in some details.Comment: 7 pages, slightly revised version, no figure
Complexity Measures from Interaction Structures
We evaluate new complexity measures on the symbolic dynamics of coupled tent
maps and cellular automata. These measures quantify complexity in terms of
-th order statistical dependencies that cannot be reduced to interactions
between units. We demonstrate that these measures are able to identify
complex dynamical regimes.Comment: 11 pages, figures improved, minor changes to the tex
Pairwise Confusion for Fine-Grained Visual Classification
Fine-Grained Visual Classification (FGVC) datasets contain small sample
sizes, along with significant intra-class variation and inter-class similarity.
While prior work has addressed intra-class variation using localization and
segmentation techniques, inter-class similarity may also affect feature
learning and reduce classification performance. In this work, we address this
problem using a novel optimization procedure for the end-to-end neural network
training on FGVC tasks. Our procedure, called Pairwise Confusion (PC) reduces
overfitting by intentionally {introducing confusion} in the activations. With
PC regularization, we obtain state-of-the-art performance on six of the most
widely-used FGVC datasets and demonstrate improved localization ability. {PC}
is easy to implement, does not need excessive hyperparameter tuning during
training, and does not add significant overhead during test time.Comment: Camera-Ready version for ECCV 201
- âŠ