47,483 research outputs found
Reduction of Markov Chains using a Value-of-Information-Based Approach
In this paper, we propose an approach to obtain reduced-order models of
Markov chains. Our approach is composed of two information-theoretic processes.
The first is a means of comparing pairs of stationary chains on different state
spaces, which is done via the negative Kullback-Leibler divergence defined on a
model joint space. Model reduction is achieved by solving a
value-of-information criterion with respect to this divergence. Optimizing the
criterion leads to a probabilistic partitioning of the states in the high-order
Markov chain. A single free parameter that emerges through the optimization
process dictates both the partition uncertainty and the number of state groups.
We provide a data-driven means of choosing the `optimal' value of this free
parameter, which sidesteps needing to a priori know the number of state groups
in an arbitrary chain.Comment: Submitted to Entrop
Optimal Kullback-Leibler Aggregation via Information Bottleneck
In this paper, we present a method for reducing a regular, discrete-time
Markov chain (DTMC) to another DTMC with a given, typically much smaller number
of states. The cost of reduction is defined as the Kullback-Leibler divergence
rate between a projection of the original process through a partition function
and a DTMC on the correspondingly partitioned state space. Finding the reduced
model with minimal cost is computationally expensive, as it requires an
exhaustive search among all state space partitions, and an exact evaluation of
the reduction cost for each candidate partition. Our approach deals with the
latter problem by minimizing an upper bound on the reduction cost instead of
minimizing the exact cost; The proposed upper bound is easy to compute and it
is tight if the original chain is lumpable with respect to the partition. Then,
we express the problem in the form of information bottleneck optimization, and
propose using the agglomerative information bottleneck algorithm for searching
a sub-optimal partition greedily, rather than exhaustively. The theory is
illustrated with examples and one application scenario in the context of
modeling bio-molecular interactions.Comment: 13 pages, 4 figure
Joint segmentation of multivariate astronomical time series : bayesian sampling with a hierarchical model
Astronomy and other sciences often face the problem of detecting and characterizing structure in two or more related time series. This paper approaches such problems using Bayesian priors to represent relationships between signals with various degrees of certainty, and not just rigid constraints. The segmentation is conducted by using a hierarchical Bayesian approach to a piecewise constant Poisson rate model. A Gibbs sampling strategy allows joint estimation of the unknown parameters and hyperparameters. Results obtained with synthetic and real photon counting data illustrate the performance of the proposed algorithm
Rank-normalization, folding, and localization: An improved for assessing convergence of MCMC
Markov chain Monte Carlo is a key computational tool in Bayesian statistics,
but it can be challenging to monitor the convergence of an iterative stochastic
algorithm. In this paper we show that the convergence diagnostic
of Gelman and Rubin (1992) has serious flaws. Traditional will
fail to correctly diagnose convergence failures when the chain has a heavy tail
or when the variance varies across the chains. In this paper we propose an
alternative rank-based diagnostic that fixes these problems. We also introduce
a collection of quantile-based local efficiency measures, along with a
practical approach for computing Monte Carlo error estimates for quantiles. We
suggest that common trace plots should be replaced with rank plots from
multiple chains. Finally, we give recommendations for how these methods should
be used in practice.Comment: Minor revision for improved clarit
Rank-normalization, folding, and localization: An improved for assessing convergence of MCMC
Markov chain Monte Carlo is a key computational tool in Bayesian statistics,
but it can be challenging to monitor the convergence of an iterative stochastic
algorithm. In this paper we show that the convergence diagnostic
of Gelman and Rubin (1992) has serious flaws. Traditional will
fail to correctly diagnose convergence failures when the chain has a heavy tail
or when the variance varies across the chains. In this paper we propose an
alternative rank-based diagnostic that fixes these problems. We also introduce
a collection of quantile-based local efficiency measures, along with a
practical approach for computing Monte Carlo error estimates for quantiles. We
suggest that common trace plots should be replaced with rank plots from
multiple chains. Finally, we give recommendations for how these methods should
be used in practice.Comment: Minor revision for improved clarit
- …