10,941 research outputs found
Rank-normalization, folding, and localization: An improved for assessing convergence of MCMC
Markov chain Monte Carlo is a key computational tool in Bayesian statistics,
but it can be challenging to monitor the convergence of an iterative stochastic
algorithm. In this paper we show that the convergence diagnostic
of Gelman and Rubin (1992) has serious flaws. Traditional will
fail to correctly diagnose convergence failures when the chain has a heavy tail
or when the variance varies across the chains. In this paper we propose an
alternative rank-based diagnostic that fixes these problems. We also introduce
a collection of quantile-based local efficiency measures, along with a
practical approach for computing Monte Carlo error estimates for quantiles. We
suggest that common trace plots should be replaced with rank plots from
multiple chains. Finally, we give recommendations for how these methods should
be used in practice.Comment: Minor revision for improved clarit
Rank-normalization, folding, and localization: An improved for assessing convergence of MCMC
Markov chain Monte Carlo is a key computational tool in Bayesian statistics,
but it can be challenging to monitor the convergence of an iterative stochastic
algorithm. In this paper we show that the convergence diagnostic
of Gelman and Rubin (1992) has serious flaws. Traditional will
fail to correctly diagnose convergence failures when the chain has a heavy tail
or when the variance varies across the chains. In this paper we propose an
alternative rank-based diagnostic that fixes these problems. We also introduce
a collection of quantile-based local efficiency measures, along with a
practical approach for computing Monte Carlo error estimates for quantiles. We
suggest that common trace plots should be replaced with rank plots from
multiple chains. Finally, we give recommendations for how these methods should
be used in practice.Comment: Minor revision for improved clarit
Optimal Kullback-Leibler Aggregation via Information Bottleneck
In this paper, we present a method for reducing a regular, discrete-time
Markov chain (DTMC) to another DTMC with a given, typically much smaller number
of states. The cost of reduction is defined as the Kullback-Leibler divergence
rate between a projection of the original process through a partition function
and a DTMC on the correspondingly partitioned state space. Finding the reduced
model with minimal cost is computationally expensive, as it requires an
exhaustive search among all state space partitions, and an exact evaluation of
the reduction cost for each candidate partition. Our approach deals with the
latter problem by minimizing an upper bound on the reduction cost instead of
minimizing the exact cost; The proposed upper bound is easy to compute and it
is tight if the original chain is lumpable with respect to the partition. Then,
we express the problem in the form of information bottleneck optimization, and
propose using the agglomerative information bottleneck algorithm for searching
a sub-optimal partition greedily, rather than exhaustively. The theory is
illustrated with examples and one application scenario in the context of
modeling bio-molecular interactions.Comment: 13 pages, 4 figure
- …