7,413 research outputs found

    A central limit theorem for temporally non-homogenous Markov chains with applications to dynamic programming

    Get PDF
    We prove a central limit theorem for a class of additive processes that arise naturally in the theory of finite horizon Markov decision problems. The main theorem generalizes a classic result of Dobrushin (1956) for temporally non-homogeneous Markov chains, and the principal innovation is that here the summands are permitted to depend on both the current state and a bounded number of future states of the chain. We show through several examples that this added flexibility gives one a direct path to asymptotic normality of the optimal total reward of finite horizon Markov decision problems. The same examples also explain why such results are not easily obtained by alternative Markovian techniques such as enlargement of the state space.Comment: 27 pages, 1 figur

    Information Geometry Approach to Parameter Estimation in Markov Chains

    Full text link
    We consider the parameter estimation of Markov chain when the unknown transition matrix belongs to an exponential family of transition matrices. Then, we show that the sample mean of the generator of the exponential family is an asymptotically efficient estimator. Further, we also define a curved exponential family of transition matrices. Using a transition matrix version of the Pythagorean theorem, we give an asymptotically efficient estimator for a curved exponential family.Comment: Appendix D is adde

    Exponentially slow transitions on a Markov chain: the frequency of Calcium Sparks

    Get PDF
    Calcium sparks in cardiac muscle cells occur when a cluster of Ca2+ channels open and release Ca2+ from an internal store. A simplified model of Ca2+ sparks has been developed to describe the dynamics of a cluster of channels, which is of the form of a continuous time Markov chain with nearest neighbour transitions and slowly varying jump functions. The chain displays metastability, whereby the probability distribution of the state of the system evolves exponentially slowly, with one of the metastable states occurring at the boundary. An asymptotic technique for analysing the Master equation (a differential-difference equation) associated with these Markov chains is developed using the WKB and projection methods. The method is used to re-derive a known result for a standard class of Markov chains displaying metastability, before being applied to the new class of Markov chains associated with the spark model. The mean first passage time between metastable states is calculated and an expression for the frequency of calcium sparks is derived. All asymptotic results are compared with Monte Carlo simulations

    Non-Reversible Parallel Tempering: a Scalable Highly Parallel MCMC Scheme

    Full text link
    Parallel tempering (PT) methods are a popular class of Markov chain Monte Carlo schemes used to sample complex high-dimensional probability distributions. They rely on a collection of NN interacting auxiliary chains targeting tempered versions of the target distribution to improve the exploration of the state-space. We provide here a new perspective on these highly parallel algorithms and their tuning by identifying and formalizing a sharp divide in the behaviour and performance of reversible versus non-reversible PT schemes. We show theoretically and empirically that a class of non-reversible PT methods dominates its reversible counterparts and identify distinct scaling limits for the non-reversible and reversible schemes, the former being a piecewise-deterministic Markov process and the latter a diffusion. These results are exploited to identify the optimal annealing schedule for non-reversible PT and to develop an iterative scheme approximating this schedule. We provide a wide range of numerical examples supporting our theoretical and methodological contributions. The proposed methodology is applicable to sample from a distribution π\pi with a density LL with respect to a reference distribution π0\pi_0 and compute the normalizing constant. A typical use case is when π0\pi_0 is a prior distribution, LL a likelihood function and π\pi the corresponding posterior.Comment: 74 pages, 30 figures. The method is implemented in an open source probabilistic programming available at https://github.com/UBC-Stat-ML/blangSD
    • …
    corecore