148 research outputs found
A comparative reliability analysis of ETCS train radio communications
StoCharts have been proposed as a UML statechart extension for performance and dependability evaluation, and were applied in the context of train radio reliability assessment to show the principal tractability of realistic cases with this approach. In this paper, we extend on this bare feasibility result in two important directions. First, we sketch the cornerstones of a mechanizable translation of StoCharts to MoDeST. The latter is a process algebra-based formalism supported by the Motor/Mƶbius tool tandem. Second, we exploit this translation for a detailed analysis of the train radio case study
Probabilistic Model-Based Safety Analysis
Model-based safety analysis approaches aim at finding critical failure
combinations by analysis of models of the whole system (i.e. software,
hardware, failure modes and environment). The advantage of these methods
compared to traditional approaches is that the analysis of the whole system
gives more precise results. Only few model-based approaches have been applied
to answer quantitative questions in safety analysis, often limited to analysis
of specific failure propagation models, limited types of failure modes or
without system dynamics and behavior, as direct quantitative analysis is uses
large amounts of computing resources. New achievements in the domain of
(probabilistic) model-checking now allow for overcoming this problem.
This paper shows how functional models based on synchronous parallel
semantics, which can be used for system design, implementation and qualitative
safety analysis, can be directly re-used for (model-based) quantitative safety
analysis. Accurate modeling of different types of probabilistic failure
occurrence is shown as well as accurate interpretation of the results of the
analysis. This allows for reliable and expressive assessment of the safety of a
system in early design stages
Sigref ā A Symbolic Bisimulation Tool Box
We present a uniform signature-based approach to compute the most popular bisimulations. Our approach is implemented symbolically using BDDs, which enables the handling of very large transition systems. Signatures for the bisimulations are built up from a few generic building blocks, which naturally correspond to efficient BDD operations. Thus, the definition of an appropriate signature is the key for a rapid development of algorithms for other types of bisimulation.
We provide experimental evidence of the viability of this approach by presenting computational results for many bisimulations on real-world instances. The experiments show cases where our framework can handle state spaces efficiently that are far too large to handle for any tool that requires an explicit state space description.
This work was partly supported by the German Research Council (DFG) as part of the Transregional Collaborative Research Center āAutomatic Verification and Analysis of Complex Systemsā (SFB/TR 14 AVACS). See www.avacs.org for more information
Probabilistic Bisimulations for PCTL Model Checking of Interval MDPs
Verification of PCTL properties of MDPs with convex uncertainties has been
investigated recently by Puggelli et al. However, model checking algorithms
typically suffer from state space explosion. In this paper, we address
probabilistic bisimulation to reduce the size of such an MDPs while preserving
PCTL properties it satisfies. We discuss different interpretations of
uncertainty in the models which are studied in the literature and that result
in two different definitions of bisimulations. We give algorithms to compute
the quotients of these bisimulations in time polynomial in the size of the
model and exponential in the uncertain branching. Finally, we show by a case
study that large models in practice can have small branching and that a
substantial state space reduction can be achieved by our approach.Comment: In Proceedings SynCoP 2014, arXiv:1403.784
Explaining Data Type Reduction in the Shape Analysis Framework
Automatic formal verification of systems composed of a large or even unbounded
number of components is difficult as the state space of these systems is
prohibitively large. Abstraction techniques automatically construct finite
approximations of infinite-state systems from which safe information about
the original system can be inferred. We study two abstraction techniques
shape analysis, a technique from program analysis, and data type reduction,
originating from model checking. Until recently we did not properly understand
how shape analysis and data type reduction relate. In this talk, we shed light
on this relation in a comprehensive way. This is a step towards a more
unified view of abstraction employed in the static analysis and model checking
community
Breaking Dense Structures: Proving Stability of Densely Structured Hybrid Systems
Abstraction and refinement is widely used in software development. Such
techniques are valuable since they allow to handle even more complex systems.
One key point is the ability to decompose a large system into subsystems,
analyze those subsystems and deduce properties of the larger system. As
cyber-physical systems tend to become more and more complex, such techniques
become more appealing.
In 2009, Oehlerking and Theel presented a (de-)composition technique for
hybrid systems. This technique is graph-based and constructs a Lyapunov
function for hybrid systems having a complex discrete state space. The
technique consists of (1) decomposing the underlying graph of the hybrid system
into subgraphs, (2) computing multiple local Lyapunov functions for the
subgraphs, and finally (3) composing the local Lyapunov functions into a
piecewise Lyapunov function. A Lyapunov function can serve multiple purposes,
e.g., it certifies stability or termination of a system or allows to construct
invariant sets, which in turn may be used to certify safety and security.
In this paper, we propose an improvement to the decomposing technique, which
relaxes the graph structure before applying the decomposition technique. Our
relaxation significantly reduces the connectivity of the graph by exploiting
super-dense switching. The relaxation makes the decomposition technique more
efficient on one hand and on the other allows to decompose a wider range of
graph structures.Comment: In Proceedings ESSS 2015, arXiv:1506.0325
Automatic Verification of Parametric Specifications with Complex Topologies
The focus of this paper is on reducing the complexity in verification by exploiting modularity at various levels: in specification, in verification, and structurally. \begin{itemize} \item For specifications, we use the modular language CSP-OZ-DC, which allows us to decouple verification tasks concerning data from those concerning durations. \item At the verification level, we exploit modularity in theorem proving for rich data structures and use this for invariant checking. \item At the structural level, we analyze possibilities for modular verification of systems consisting of various components which interact. \end{itemize} We illustrate these ideas by automatically verifying safety properties of a case study from the European Train Control System standard, which extends previous examples by comprising a complex track topology with lists of track segments and trains with different routes
- ā¦