41 research outputs found
StocHy: automated verification and synthesis of stochastic processes
StocHy is a software tool for the quantitative analysis of discrete-time
stochastic hybrid systems (SHS). StocHy accepts a high-level description of
stochastic models and constructs an equivalent SHS model. The tool allows to
(i) simulate the SHS evolution over a given time horizon; and to automatically
construct formal abstractions of the SHS. Abstractions are then employed for
(ii) formal verification or (iii) control (policy, strategy) synthesis. StocHy
allows for modular modelling, and has separate simulation, verification and
synthesis engines, which are implemented as independent libraries. This allows
for libraries to be easily used and for extensions to be easily built. The tool
is implemented in C++ and employs manipulations based on vector calculus, the
use of sparse matrices, the symbolic construction of probabilistic kernels, and
multi-threading. Experiments show StocHy's markedly improved performance when
compared to existing abstraction-based approaches: in particular, StocHy beats
state-of-the-art tools in terms of precision (abstraction error) and
computational effort, and finally attains scalability to large-sized models (12
continuous dimensions). StocHy is available at www.gitlab.com/natchi92/StocHy
Distribution-based bisimulation for labelled Markov processes
In this paper we propose a (sub)distribution-based bisimulation for labelled
Markov processes and compare it with earlier definitions of state and event
bisimulation, which both only compare states. In contrast to those state-based
bisimulations, our distribution bisimulation is weaker, but corresponds more
closely to linear properties. We construct a logic and a metric to describe our
distribution bisimulation and discuss linearity, continuity and compositional
properties.Comment: Accepted by FORMATS 201
Probabilistic Timed Automata with Clock-Dependent Probabilities
Probabilistic timed automata are classical timed automata extended with
discrete probability distributions over edges. We introduce clock-dependent
probabilistic timed automata, a variant of probabilistic timed automata in
which transition probabilities can depend linearly on clock values.
Clock-dependent probabilistic timed automata allow the modelling of a
continuous relationship between time passage and the likelihood of system
events. We show that the problem of deciding whether the maximum probability of
reaching a certain location is above a threshold is undecidable for
clock-dependent probabilistic timed automata. On the other hand, we show that
the maximum and minimum probability of reaching a certain location in
clock-dependent probabilistic timed automata can be approximated using a
region-graph-based approach.Comment: Full version of a paper published at RP 201
Quantitative Approximation of the Probability Distribution of a Markov Process by Formal Abstractions
The goal of this work is to formally abstract a Markov process evolving in
discrete time over a general state space as a finite-state Markov chain, with
the objective of precisely approximating its state probability distribution in
time, which allows for its approximate, faster computation by that of the
Markov chain. The approach is based on formal abstractions and employs an
arbitrary finite partition of the state space of the Markov process, and the
computation of average transition probabilities between partition sets. The
abstraction technique is formal, in that it comes with guarantees on the
introduced approximation that depend on the diameters of the partitions: as
such, they can be tuned at will. Further in the case of Markov processes with
unbounded state spaces, a procedure for precisely truncating the state space
within a compact set is provided, together with an error bound that depends on
the asymptotic properties of the transition kernel of the original process. The
overall abstraction algorithm, which practically hinges on piecewise constant
approximations of the density functions of the Markov process, is extended to
higher-order function approximations: these can lead to improved error bounds
and associated lower computational requirements. The approach is practically
tested to compute probabilistic invariance of the Markov process under study,
and is compared to a known alternative approach from the literature.Comment: 29 pages, Journal of Logical Methods in Computer Scienc