127,714 research outputs found
Quantum Markov fields on graphs
We introduce generalized quantum Markov states and generalized d-Markov
chains which extend the notion quantum Markov chains on spin systems to that on
-algebras defined by general graphs. As examples of generalized d-Markov
chains, we construct the entangled Markov fields on tree graphs. The concrete
examples of generalized d-Markov chains on Cayley trees are also investigated.Comment: 23 pages, 1 figure. accepted to "Infinite Dimensional Anal. Quantum
Probability & Related Topics
Model checking for imprecise Markov chains.
We extend probabilistic computational tree logic for expressing properties of Markov chains to imprecise Markov chains, and provide an efficient algorithm for model checking of imprecise Markov chains. Thereby, we provide a formal framework to answer a very wide range of questions about imprecise Markov chains, in a systematic and computationally efficient way
Qualitative Reachability for Open Interval Markov Chains
Interval Markov chains extend classical Markov chains with the possibility to
describe transition probabilities using intervals, rather than exact values.
While the standard formulation of interval Markov chains features closed
intervals, previous work has considered also open interval Markov chains, in
which the intervals can also be open or half-open. In this paper we focus on
qualitative reachability problems for open interval Markov chains, which
consider whether the optimal (maximum or minimum) probability with which a
certain set of states can be reached is equal to 0 or 1. We present
polynomial-time algorithms for these problems for both of the standard
semantics of interval Markov chains. Our methods do not rely on the closure of
open intervals, in contrast to previous approaches for open interval Markov
chains, and can characterise situations in which probability 0 or 1 can be
attained not exactly but arbitrarily closely.Comment: Full version of a paper published at RP 201
Hitting times and probabilities for imprecise Markov chains
We consider the problem of characterising expected hitting times and hitting probabilities for imprecise Markov chains. To this end, we consider three distinct ways in which imprecise Markov chains have been defined in the literature: as sets of homogeneous Markov chains, as sets of more general stochastic processes, and as game-theoretic probability models. Our first contribution is that all these different types of imprecise Markov chains have the same lower and upper expected hitting times, and similarly the hitting probabilities are the same for these three types. Moreover, we provide a characterisation of these quantities that directly generalises a similar characterisation for precise, homogeneous Markov chains
Hitting Times and Probabilities for Imprecise Markov Chains
We consider the problem of characterising expected hitting times and hitting
probabilities for imprecise Markov chains. To this end, we consider three
distinct ways in which imprecise Markov chains have been defined in the
literature: as sets of homogeneous Markov chains, as sets of more general
stochastic processes, and as game-theoretic probability models. Our first
contribution is that all these different types of imprecise Markov chains have
the same lower and upper expected hitting times, and similarly the hitting
probabilities are the same for these three types. Moreover, we provide a
characterisation of these quantities that directly generalises a similar
characterisation for precise, homogeneous Markov chains
Decisive Markov Chains
We consider qualitative and quantitative verification problems for
infinite-state Markov chains. We call a Markov chain decisive w.r.t. a given
set of target states F if it almost certainly eventually reaches either F or a
state from which F can no longer be reached. While all finite Markov chains are
trivially decisive (for every set F), this also holds for many classes of
infinite Markov chains. Infinite Markov chains which contain a finite attractor
are decisive w.r.t. every set F. In particular, this holds for probabilistic
lossy channel systems (PLCS). Furthermore, all globally coarse Markov chains
are decisive. This class includes probabilistic vector addition systems (PVASS)
and probabilistic noisy Turing machines (PNTM). We consider both safety and
liveness problems for decisive Markov chains, i.e., the probabilities that a
given set of states F is eventually reached or reached infinitely often,
respectively. 1. We express the qualitative problems in abstract terms for
decisive Markov chains, and show an almost complete picture of its decidability
for PLCS, PVASS and PNTM. 2. We also show that the path enumeration algorithm
of Iyer and Narasimha terminates for decisive Markov chains and can thus be
used to solve the approximate quantitative safety problem. A modified variant
of this algorithm solves the approximate quantitative liveness problem. 3.
Finally, we show that the exact probability of (repeatedly) reaching F cannot
be effectively expressed (in a uniform way) in Tarski-algebra for either PLCS,
PVASS or (P)NTM.Comment: 32 pages, 0 figure
- …