18 research outputs found

    Practical applications of probabilistic model checking to communication protocols

    Get PDF
    Probabilistic model checking is a formal verification technique for the analysis of systems that exhibit stochastic behaviour. It has been successfully employed in an extremely wide array of application domains including, for example, communication and multimedia protocols, security and power management. In this chapter we focus on the applicability of these techniques to the analysis of communication protocols. An analysis of the performance of such systems must successfully incorporate several crucial aspects, including concurrency between multiple components, real-time constraints and randomisation. Probabilistic model checking, in particular using probabilistic timed automata, is well suited to such an analysis. We provide an overview of this area, with emphasis on an industrially relevant case study: the IEEE 802.3 (CSMA/CD) protocol. We also discuss two contrasting approaches to the implementation of probabilistic model checking, namely those based on numerical computation and those based on discrete-event simulation. Using results from the two tools PRISM and APMC, we summarise the advantages, disadvantages and trade-offs associated with these techniques

    Pathobiology of Anaplastic Large Cell Lymphoma

    Get PDF
    The authors revise the concept of anaplastic large cell lymphoma (ALCL) in the light of the recently updated WHO classification of Tumors of Hematopoietic and Lymphoid Tissues both on biological and clinical grounds. The main histological findings are illustrated with special reference to the cytological spectrum that is indeed characteristic of the tumor. The phenotype is reported in detail: the expression of the ALK protein as well as the chromosomal abnormalities is discussed with their potential pathogenetic implications. The clinical features of ALCL are presented by underlining the difference in terms of response to therapy and survival between the ALK-positive and ALK-negative forms. Finally, the biological rationale for potential innovative targeted therapies is presented

    Modelling dynamic reliability via Fluid Petri Nets

    Get PDF
    Combinatorial models for reliability analysis (like fault-trees or block diagram) are static models that cannot include any type of component dependence. In the CTMC (Continuous Time Markov Chain) framework, the transition rates can depend on the state of the system thus allowing the analyst to include some dependencies among components. However, in more general terms, the system reliability may depend on parameters or quantities that vary continuously in time (like temperature, pressure, distance, etc.). Systems whose behavior in time can be described by discrete as well as continuous variables, are called hybrid systems. In the dependability literature, the case in which the reliability characteristics vary continuously versus a process parameter, is sometimes referred to as dynamic reliability [1]. The modelling and analysis of hybrid dynamic systems is an open research area. The present paper discusses the evaluation of a benchmark on dynamic reliability proposed in [1] via a modelling framework called Fluid Stochastic Petri Net (FSPN)

    The Conversion of Dynamic Fault Trees to Stochastic Petri Nets, as a case of Graph Transformation

    Get PDF
    AbstractA model-to-model transformation from Dynamic Fault Trees to Stochastic Petri Nets, by means of graph transformation rules, is presented in this paper. Dynamic Fault Trees (DFT) are used for the reliability analysis of complex and large systems and represent by means of gates, how combinations or sequences of component failure events, lead to the failure of the system. DFTs need the state space solution which can be obtained by converting a DFT to a Stochastic Petri Net: this task is expressed by means of graph transformation rules, and is applied to a case of system

    Backward Bisimulation in Markov Chain Model Checking

    Full text link

    Fast algorithm for real-time rings reconstruction

    Get PDF
    The GAP project is dedicated to study the application of GPU in several contexts in which real-time response is important to take decisions. The definition of real-time depends on the application under study, ranging from answer time of μs up to several hours in case of very computing intensive task. During this conference we presented our work in low level triggers [1] [2] and high level triggers [3] in high energy physics experiments, and specific application for nuclear magnetic resonance (NMR) [4] [5] and cone-beam CT [6]. Apart from the study of dedicated solution to decrease the latency due to data transport and preparation, the computing algorithms play an essential role in any GPU application. In this contribution, we show an original algorithm developed for triggers application, to accelerate the ring reconstruction in RICH detector when it is not possible to have seeds for reconstruction from external trackers

    ECLAP 2012 Conference on Information Technologies for Performing Arts, Media Access and Entertainment

    Get PDF
    It has been a long history of Information Technology innovations within the Cultural Heritage areas. The Performing arts has also been enforced with a number of new innovations which unveil a range of synergies and possibilities. Most of the technologies and innovations produced for digital libraries, media entertainment and education can be exploited in the field of performing arts, with adaptation and repurposing. Performing arts offer many interesting challenges and opportunities for research and innovations and exploitation of cutting edge research results from interdisciplinary areas. For these reasons, the ECLAP 2012 can be regarded as a continuation of past conferences such as AXMEDIS and WEDELMUSIC (both pressed by IEEE and FUP). ECLAP is an European Commission project to create a social network and media access service for performing arts institutions in Europe, to create the e-library of performing arts, exploiting innovative solutions coming from the ICT

    Performance Regression Detection in DevOps

    Get PDF
    Performance is an important aspect of software quality. The goals of performance are typically defined by setting upper and lower bounds for response time and throughput of a system and physical level measurements such as CPU, memory, and I/O. To meet such performance goals, several performance-related activities are needed in development (Dev) and operations (Ops). Large software system failures are often due to performance issues rather than functional bugs. One of the most important performance issues is performance regression. Although performance regressions are not all bugs, they often have a direct impact on users’ experience of the system. The process of detection of performance regressions in development and operations is faced with challenges. First, the detection of performance regression is conducted after the fact, i.e., after the system is built and deployed in the field or dedicated performance testing environments. Large amounts of resources are required to detect, locate, understand, and fix performance regressions at such a late stage in the development cycle. Second, even we can detect a performance regression, it is extremely hard to fix it because other changes are applied to the system after the introduction of the regression. These challenges call for further in-depth analyses of the performance regression. In this thesis, to avoid performance regression slipping into operation, we first perform an exploratory study on the source code changes that introduce performance regressions in order to understand root-causes of performance regression in the source code level. Second, we propose an approach that automatically predicts whether a test would manifest performance regressions in a code commit. Most of the performance issues are related to configurations. Therefore, third, we propose an approach that predicts whether a configuration option manifests a performance variation issue. To assist practitioners to analyze system performance with operational data, we propose an approach to recovering field-representative workload that can be used to detect performance regression
    corecore