3,198 research outputs found
Novel Lower Bounds on the Entropy Rate of Binary Hidden Markov Processes
Recently, Samorodnitsky proved a strengthened version of Mrs. Gerber's Lemma,
where the output entropy of a binary symmetric channel is bounded in terms of
the average entropy of the input projected on a random subset of coordinates.
Here, this result is applied for deriving novel lower bounds on the entropy
rate of binary hidden Markov processes. For symmetric underlying Markov
processes, our bound improves upon the best known bound in the very noisy
regime. The nonsymmetric case is also considered, and explicit bounds are
derived for Markov processes that satisfy the -RLL constraint
Estimating ensemble flows on a hidden Markov chain
We propose a new framework to estimate the evolution of an ensemble of
indistinguishable agents on a hidden Markov chain using only aggregate output
data. This work can be viewed as an extension of the recent developments in
optimal mass transport and Schr\"odinger bridges to the finite state space
hidden Markov chain setting. The flow of the ensemble is estimated by solving a
maximum likelihood problem, which has a convex formulation at the
infinite-particle limit, and we develop a fast numerical algorithm for it. We
illustrate in two numerical examples how this framework can be used to track
the flow of identical and indistinguishable dynamical systems.Comment: 8 pages, 4 figure
Taylor series expansions for the entropy rate of Hidden Markov Processes
Finding the entropy rate of Hidden Markov Processes is an active research
topic, of both theoretical and practical importance. A recently used approach
is studying the asymptotic behavior of the entropy rate in various regimes. In
this paper we generalize and prove a previous conjecture relating the entropy
rate to entropies of finite systems. Building on our new theorems, we establish
series expansions for the entropy rate in two different regimes. We also study
the radius of convergence of the two series expansions
Prediction and Generation of Binary Markov Processes: Can a Finite-State Fox Catch a Markov Mouse?
Understanding the generative mechanism of a natural system is a vital
component of the scientific method. Here, we investigate one of the fundamental
steps toward this goal by presenting the minimal generator of an arbitrary
binary Markov process. This is a class of processes whose predictive model is
well known. Surprisingly, the generative model requires three distinct
topologies for different regions of parameter space. We show that a previously
proposed generator for a particular set of binary Markov processes is, in fact,
not minimal. Our results shed the first quantitative light on the relative
(minimal) costs of prediction and generation. We find, for instance, that the
difference between prediction and generation is maximized when the process is
approximately independently, identically distributed.Comment: 12 pages, 12 figures;
http://csc.ucdavis.edu/~cmg/compmech/pubs/gmc.ht
On the Performance of Short Block Codes over Finite-State Channels in the Rare-Transition Regime
As the mobile application landscape expands, wireless networks are tasked
with supporting different connection profiles, including real-time traffic and
delay-sensitive communications. Among many ensuing engineering challenges is
the need to better understand the fundamental limits of forward error
correction in non-asymptotic regimes. This article characterizes the
performance of random block codes over finite-state channels and evaluates
their queueing performance under maximum-likelihood decoding. In particular,
classical results from information theory are revisited in the context of
channels with rare transitions, and bounds on the probabilities of decoding
failure are derived for random codes. This creates an analysis framework where
channel dependencies within and across codewords are preserved. Such results
are subsequently integrated into a queueing problem formulation. For instance,
it is shown that, for random coding on the Gilbert-Elliott channel, the
performance analysis based on upper bounds on error probability provides very
good estimates of system performance and optimum code parameters. Overall, this
study offers new insights about the impact of channel correlation on the
performance of delay-aware, point-to-point communication links. It also
provides novel guidelines on how to select code rates and block lengths for
real-time traffic over wireless communication infrastructures
Hybrid modeling, HMM/NN architectures, and protein applications
We describe a hybrid modeling approach where the parameters of a model are calculated and modulated by another model, typically a neural network (NN), to avoid both overfitting and underfitting. We develop the approach for the case of Hidden Markov Models (HMMs), by deriving a class of hybrid HMM/NN architectures. These architectures can be trained with unified algorithms that blend HMM dynamic programming with NN backpropagation. In the case of complex data, mixtures of HMMs or modulated HMMs must be used. NNs can then be applied both to the parameters of each single HMM, and to the switching or modulation of the models, as a function of input or context. Hybrid HMM/NN architectures provide a flexible NN parameterization for the control of model structure and complexity. At the same time, they can capture distributions that, in practice, are inaccessible to single HMMs. The HMM/NN hybrid approach is tested, in its simplest form, by constructing a model of the immunoglobulin protein family. A hybrid model is trained, and a multiple alignment derived, with less than a fourth of the number of parameters used with previous single HMMs
Localizing the Latent Structure Canonical Uncertainty: Entropy Profiles for Hidden Markov Models
This report addresses state inference for hidden Markov models. These models
rely on unobserved states, which often have a meaningful interpretation. This
makes it necessary to develop diagnostic tools for quantification of state
uncertainty. The entropy of the state sequence that explains an observed
sequence for a given hidden Markov chain model can be considered as the
canonical measure of state sequence uncertainty. This canonical measure of
state sequence uncertainty is not reflected by the classic multivariate state
profiles computed by the smoothing algorithm, which summarizes the possible
state sequences. Here, we introduce a new type of profiles which have the
following properties: (i) these profiles of conditional entropies are a
decomposition of the canonical measure of state sequence uncertainty along the
sequence and makes it possible to localize this uncertainty, (ii) these
profiles are univariate and thus remain easily interpretable on tree
structures. We show how to extend the smoothing algorithms for hidden Markov
chain and tree models to compute these entropy profiles efficiently.Comment: Submitted to Journal of Machine Learning Research; No RR-7896 (2012
Identifying Functional Thermodynamics in Autonomous Maxwellian Ratchets
We introduce a family of Maxwellian Demons for which correlations among
information bearing degrees of freedom can be calculated exactly and in compact
analytical form. This allows one to precisely determine Demon functional
thermodynamic operating regimes, when previous methods either misclassify or
simply fail due to approximations they invoke. This reveals that these Demons
are more functional than previous candidates. They too behave either as
engines, lifting a mass against gravity by extracting energy from a single heat
reservoir, or as Landauer erasers, consuming external work to remove
information from a sequence of binary symbols by decreasing their individual
uncertainty. Going beyond these, our Demon exhibits a new functionality that
erases bits not by simply decreasing individual-symbol uncertainty, but by
increasing inter-bit correlations (that is, by adding temporal order) while
increasing single-symbol uncertainty. In all cases, but especially in the new
erasure regime, exactly accounting for informational correlations leads to
tight bounds on Demon performance, expressed as a refined Second Law of
Thermodynamics that relies on the Kolmogorov-Sinai entropy for dynamical
processes and not on changes purely in system configurational entropy, as
previously employed. We rigorously derive the refined Second Law under minimal
assumptions and so it applies quite broadly---for Demons with and without
memory and input sequences that are correlated or not. We note that general
Maxwellian Demons readily violate previously proposed, alternative such bounds,
while the current bound still holds.Comment: 13 pages, 9 figures,
http://csc.ucdavis.edu/~cmg/compmech/pubs/mrd.ht
A relative entropy rate method for path space sensitivity analysis of stationary complex stochastic dynamics
We propose a new sensitivity analysis methodology for complex stochastic
dynamics based on the Relative Entropy Rate. The method becomes computationally
feasible at the stationary regime of the process and involves the calculation
of suitable observables in path space for the Relative Entropy Rate and the
corresponding Fisher Information Matrix. The stationary regime is crucial for
stochastic dynamics and here allows us to address the sensitivity analysis of
complex systems, including examples of processes with complex landscapes that
exhibit metastability, non-reversible systems from a statistical mechanics
perspective, and high-dimensional, spatially distributed models. All these
systems exhibit, typically non-gaussian stationary probability distributions,
while in the case of high-dimensionality, histograms are impossible to
construct directly. Our proposed methods bypass these challenges relying on the
direct Monte Carlo simulation of rigorously derived observables for the
Relative Entropy Rate and Fisher Information in path space rather than on the
stationary probability distribution itself. We demonstrate the capabilities of
the proposed methodology by focusing here on two classes of problems: (a)
Langevin particle systems with either reversible (gradient) or non-reversible
(non-gradient) forcing, highlighting the ability of the method to carry out
sensitivity analysis in non-equilibrium systems; and, (b) spatially extended
Kinetic Monte Carlo models, showing that the method can handle high-dimensional
problems
- …