29 research outputs found
Probabilistic Process Algebra
Every day we witness the fast development of the hardware and software technology. This, of course, is the reason that new and more complex systems controlled by some kind of computational-based devices become an unseparated part of our daily life. As more as the system complexity increases, as more the reasoning about its correct behaviour becomes dif??cult. A variety of consequences may occur as a result of a failure, ranging from simple annoying to life threatening ones. Thus for some systems it is crucial that they exhibit a correct functioning. However, for systems with an extremely complex construction it is almost impossible to give an absolute guarantee for their correctness. In this case, it is still satisfactory to know that the possibility for a system to fail is low enough. Formal methods have been developed for establishing correctness of computer systems. They provide rigorous methods with which one can formally specify properties of a systems's intended behaviour, and also can check if the system conforms to that speci??cation. In case of complex systems we need a formal method that allows us to reason in compositional way, it provides us with techniques that can be used to build larger systems from the composition of smaller ones. Process algebra carries exactly this idea; it provides operators that allow to compose processes in order to obtain a more complex process. Besides, every process algebra contains a set of axioms. Every axiom is an algebraic equation that carries our intuition and insight in process behaviour, it expresses which two processes behaviour we consider equal. In such a way, manipulation with processes becomes manipulation with equations in the algebraic sense. But, equations and operators do not have any meaning unless we place them in a certain real ¿world¿ and match the terms of the process algebra with the entities of the real world. This step is traditionally called ¿giving a semantic of the syntax¿. The structure constructed in this way is called a model of the considered process algebra. For every given process algebra we can construct an in??nite number of models, but only several of them are interesting for the purpose process algebra was developed as a formal method. However, there is a tendency always to use so-called a bisimulation model. In this thesis we propose several process algebras and construct their models based on the notion of bisimulation
Quantitative Safety: Linking Proof-Based Verification with Model Checking for Probabilistic Systems
This paper presents a novel approach for augmenting proof-based verification
with performance-style analysis of the kind employed in state-of-the-art model
checking tools for probabilistic systems. Quantitative safety properties
usually specified as probabilistic system invariants and modeled in proof-based
environments are evaluated using bounded model checking techniques.
Our specific contributions include the statement of a theorem that is central
to model checking safety properties of proof-based systems, the establishment
of a procedure; and its full implementation in a prototype system (YAGA) which
readily transforms a probabilistic model specified in a proof-based environment
to its equivalent verifiable PRISM model equipped with reward structures. The
reward structures capture the exact interpretation of the probabilistic
invariants and can reveal succinct information about the model during
experimental investigations. Finally, we demonstrate the novelty of the
technique on a probabilistic library case study
Markovian Testing Equivalence and Exponentially Timed Internal Actions
In the theory of testing for Markovian processes developed so far,
exponentially timed internal actions are not admitted within processes. When
present, these actions cannot be abstracted away, because their execution takes
a nonzero amount of time and hence can be observed. On the other hand, they
must be carefully taken into account, in order not to equate processes that are
distinguishable from a timing viewpoint. In this paper, we recast the
definition of Markovian testing equivalence in the framework of a Markovian
process calculus including exponentially timed internal actions. Then, we show
that the resulting behavioral equivalence is a congruence, has a sound and
complete axiomatization, has a modal logic characterization, and can be decided
in polynomial time
Branching bisimulation for probabilistic systems: characteristics and decidability
We address the concept of abstraction in the setting of probabilistic reactive systems, and study its formal underpinnings for the strictly alternating model of Hansson. In particular, we define the notion of branching bisimilarity and study its properties by studying two other equivalence relations, viz. coloured trace equivalence and branching bisimilarity using maximal probabilities. We show that both alternatives coincide with branching bisimilarity. The alternative characterisations have their own merits and focus on different aspects of branching bisimilarity. Coloured trace equivalence can be understood without knowledge of probability theory and is independent of the notion of a scheduler. Branching bisimilarity, rephrased in terms of maximal probabilities gives rise to an algorithm of polynomial complexity for deciding the equivalence. Together they give a better understanding of branching bisimilarity. Furthermore, we show that the notions of branching bisimilarity in the alternating model of Hansson and in the nonalternating model of Segala differ: branching bisimilarity in the latter setting turns out to discriminate between systems that are intuitively branching bisimilar
Towards reduction of Paradigm coordination models
The coordination modelling language Paradigm addresses collaboration between
components in terms of dynamic constraints. Within a Paradigm model, component
dynamics are consistently specified at a detailed and a global level of
abstraction. To enable automated verification of Paradigm models, a translation
of Paradigm into process algebra has been defined in previous work. In this
paper we investigate, guided by a client-server example, reduction of Paradigm
models based on a notion of global inertness. Representation of Paradigm models
as process algebraic specifications helps to establish a property-preserving
equivalence relation between the original and the reduced Paradigm model.
Experiments indicate that in this way larger Paradigm models can be analyzed.Comment: In Proceedings PACO 2011, arXiv:1108.145
Verifying Real-Time Systems using Explicit-time Description Methods
Timed model checking has been extensively researched in recent years. Many
new formalisms with time extensions and tools based on them have been
presented. On the other hand, Explicit-Time Description Methods aim to verify
real-time systems with general untimed model checkers. Lamport presented an
explicit-time description method using a clock-ticking process (Tick) to
simulate the passage of time together with a group of global variables for time
requirements. This paper proposes a new explicit-time description method with
no reliance on global variables. Instead, it uses rendezvous synchronization
steps between the Tick process and each system process to simulate time. This
new method achieves better modularity and facilitates usage of more complex
timing constraints. The two explicit-time description methods are implemented
in DIVINE, a well-known distributed-memory model checker. Preliminary
experiment results show that our new method, with better modularity, is
comparable to Lamport's method with respect to time and memory efficiency
A framework for compositional verification of security protocols
Automatic security protocol analysis is currently feasible only for small protocols. Since larger protocols quite often are composed of many small protocols, compositional analysis is an attractive, but non-trivial approach.
We have developed a framework for compositional analysis of a large class of security protocols. The framework is intended to facilitate automatic as well as manual verification of large structured security protocols. Our approach is to verify properties of component protocols in a multi-protocol environment, then deduce properties about the composed protocol. To reduce the complexity of multi-protocol verification, we introduce a notion of protocol independence and prove a number of theorems that enable analysis of independent component protocols in isolation.
To illustrate the applicability of our framework to real-world protocols, we study a key establishment sequence in WiMAX consisting of three subprotocols. Except for a small amount of trivial reasoning, the analysis is done using automatic tools
Modelling Clock Synchronization in the Chess gMAC WSN Protocol
We present a detailled timed automata model of the clock synchronization
algorithm that is currently being used in a wireless sensor network (WSN) that
has been developed by the Dutch company Chess. Using the Uppaal model checker,
we establish that in certain cases a static, fully synchronized network may
eventually become unsynchronized if the current algorithm is used, even in a
setting with infinitesimal clock drifts
Prototyping the Semantics of a DSL using ASF+SDF: Link to Formal Verification of DSL Models
A formal definition of the semantics of a domain-specific language (DSL) is a
key prerequisite for the verification of the correctness of models specified
using such a DSL and of transformations applied to these models. For this
reason, we implemented a prototype of the semantics of a DSL for the
specification of systems consisting of concurrent, communicating objects. Using
this prototype, models specified in the DSL can be transformed to labeled
transition systems (LTS). This approach of transforming models to LTSs allows
us to apply existing tools for visualization and verification to models with
little or no further effort. The prototype is implemented using the ASF+SDF
Meta-Environment, an IDE for the algebraic specification language ASF+SDF,
which offers efficient execution of the transformation as well as the ability
to read models and produce LTSs without any additional pre or post processing.Comment: In Proceedings AMMSE 2011, arXiv:1106.596