45 research outputs found

    Quantitative Verification: Formal Guarantees for Timeliness, Reliability and Performance

    Get PDF
    Computerised systems appear in almost all aspects of our daily lives, often in safety-critical scenarios such as embedded control systems in cars and aircraft or medical devices such as pacemakers and sensors. We are thus increasingly reliant on these systems working correctly, despite often operating in unpredictable or unreliable environments. Designers of such devices need ways to guarantee that they will operate in a reliable and efficient manner. Quantitative verification is a technique for analysing quantitative aspects of a system's design, such as timeliness, reliability or performance. It applies formal methods, based on a rigorous analysis of a mathematical model of the system, to automatically prove certain precisely specified properties, e.g. ``the airbag will always deploy within 20 milliseconds after a crash'' or ``the probability of both sensors failing simultaneously is less than 0.001''. The ability to formally guarantee quantitative properties of this kind is beneficial across a wide range of application domains. For example, in safety-critical systems, it may be essential to establish credible bounds on the probability with which certain failures or combinations of failures can occur. In embedded control systems, it is often important to comply with strict constraints on timing or resources. More generally, being able to derive guarantees on precisely specified levels of performance or efficiency is a valuable tool in the design of, for example, wireless networking protocols, robotic systems or power management algorithms, to name but a few. This report gives a short introduction to quantitative verification, focusing in particular on a widely used technique called model checking, and its generalisation to the analysis of quantitative aspects of a system such as timing, probabilistic behaviour or resource usage. The intended audience is industrial designers and developers of systems such as those highlighted above who could benefit from the application of quantitative verification,but lack expertise in formal verification or modelling

    Performance Modelling and Verification of Cloud-based Auto-Scaling Policies

    Get PDF

    Towards a new methodology for design, modelling, and verification of reconfigurable distributed control systems based on a new extension to the IEC 61499 standard

    Get PDF
    In order to meet user requirements and system environment changes, reconfigurable control systems must dynamically adapt their structure and behaviour without disrupting system operation. IEC 61499 standard provides limited support for the design and verification of such systems. In fact, handling different reconfiguration scenarios at runtime is difficult since function blocks in IEC 61499 cannot be changed at run-time. Hence, this thesis promotes an IEC 61499 extension called reconfigurable function block (RFB) that increases design readability and smoothly switches to the most appropriate behaviour when a reconfiguration event occurs. To ensure system feasibility after reconfiguration, in addition to the qualitative verification, quantitative verification based on probabilistic model checking is addressed in a new RFBA approach. The latter aims to transform the designed RFB model automatically into a generalised reconfigurable timed net condition/event system model (GRTNCES) using a newly developed environment called RFBTool. The GR-TNCES fits well with RFB and preserves its semantic. Using the probabilistic model checker PRISM, the generated GR-TNCES model is checked using defined properties specified in computation tree logic. As a result, an evaluation of system performance and an estimation of reconfiguration risks are obtained. The RFBA methodology is applied on a distributed power system case study.Dynamische Anforderungen und Umgebungen erfordern rekonfigurierbare Anlagen und Steuerungssysteme. Rekonfiguration ermöglicht es einem System, seine Struktur und sein Verhalten an interne oder externe Änderungen anzupassen. Die Norm IEC 61499 wurde entwickelt, um (verteilte) Steuerungssysteme auf Basis von Funktionsbausteinen zu entwickeln. Sie bietet jedoch wenig Unterstützung für Entwurf und Verifikation. Die Tatsache, dass eine Rekonfiguration das System-Ausführungsmodell verändert, erschwert die Entwicklung in IEC 61499 zusätzlich. Daher schlägt diese Dissertation rekonfigurierbare Funktionsbausteine (RFBs) als Erweiterung der Norm vor. Ein RFB verarbeitet über einen Master-Slave-Automaten Rekonfigurationsereignisse und löst das entsprechende Verhalten aus. Diese Hierarchie trennt das Rekonfigurationsmodell vom Steuerungsmodell und vereinfacht so den Entwurf. Die Funktionalität des Entwurfs muss verifiziert werden, damit die Ausführbarkeit des Systems nach einer Rekonfiguration gewährleistet ist. Hierzu wird das entworfene RFB-Modell automatisch in ein generalised reconfigurable timed net condition/event system übersetzt. Dieses wird mit dem Model-Checker PRISM auf qualitative und quantitative Eigenschaften überprüft. Somit wird eine Bewertung der Systemperformanz und eine Einschätzung der Rekonfigurationsrisiken erreicht. Die RFB-Methodik wurde in einem Softwarewerkzeug umgesetzt und in einer Fallstudie auf ein dezentrales Stromnetz angewendet

    기계 학습을 통한 가상화 플랫폼의 라이브 마이그레이션 성능 예측

    Get PDF
    학위논문 (석사)-- 서울대학교 대학원 : 전기·컴퓨터공학부, 2016. 2. Bernhard Egger.Virtualization is a widely used technology these days as most of server computing environments are rapidly shifting to cloud computing. Live migration, one of the most compelling features in system virtualization, has been an active area of research. Attempts to predict migration performance were made, but most of those were limited to analytical approaches with relatively unstable prediction errors or not easy to extend to realistic environments as more parameters are identified and considered. In this thesis, a novel data driven approach based on the support vector regression method providing flexibility and extensibility in parameter selection is introduced to predict performance metrics such as total migration time, downtime and the total amount of transferred data, especially on QEMU which is hardware virtualization platform that is open-source and the method of this thesis is easy to adapt to various purposes. It will facilitate automated system administration with live migration more efficiently.Chapter 1 Introduction 1 Chapter 2 Background and related work 4 2.1 Live migration algorithms 5 2.2 Performance metrics 9 2.3 Existing models and evaluation attempts 11 Chapter 3 Empirical Evaluation 13 3.1 Sample generation and evaluation 13 3.2 Workloads 14 Chapter 4 Data driven approach 23 4.1 Parameter selection and migration algorithms 23 4.2 Prediction using support vector regression 24 4.3 Tool architecture 26 4.4 Single vs. multiple predictors 27 Chapter 5 Experimental evaluation 29 5.1 Training setup 29 5.2 Prediction results 30 Chapter 6 Conclusion 37 Bibliography 38 Abstract in Korean 41Maste

    Dynamic resource provisioning and secured file sharing using virtualization in cloud azure

    Get PDF
    Virtual machines (VMs) are preferred by the majority of organizations due to their high performance. VMs allow for reduced overhead with multiple systems running from the same console at the same time. A physical server is a bare-metal system whose hardware is controlled by the host operating system. A physical server runs on a single instance of OS and application. A virtual server or virtual machine encapsulates the underlying hardware and networking resources. With the existing physical server, it is difficult to migrate the tasks from one platform to another platform or to a datacentre. Centralized security is difficult to setup. But with Hypervisor the virtual machine can be deployed, for instance, with automation. Virtualization cost increases as well as a decrease in hardware and infrastructure space costs. We propose an efficient Azure cloud framework for the utilization of physical server resources at remote VM servers. The proposed framework is implemented in two phases first by integrating physical servers into virtual ones by creating virtual machines, and then by integrating virtual servers into cloud service providers in a cost-effective manner. We create a virtual network in the Azure datacenter using the local host physical server to set up the various virtual machines. Two virtual machine instances, VM1 and VM2, are created using Microsoft Hyper-V with the server Windows 2016 R. The desktop application is deployed and VM performance is monitored using the PowerShell script. Tableau is used to evaluate the physical server functionality of the worksheet for the deployed application. The proposed Physical to Virtual to Cloud model (P2V2C) model is being tested, and the performance result shows that P2V2C migration is more successful in dynamic provisioning than direct migration to cloud platform infrastructure. The research work was carried out in a secure way through the migration process from P2V2C.Web of Science111art. no. 4

    Computer Aided Verification

    Get PDF
    This open access two-volume set LNCS 10980 and 10981 constitutes the refereed proceedings of the 30th International Conference on Computer Aided Verification, CAV 2018, held in Oxford, UK, in July 2018. The 52 full and 13 tool papers presented together with 3 invited papers and 2 tutorials were carefully reviewed and selected from 215 submissions. The papers cover a wide range of topics and techniques, from algorithmic and logical foundations of verification to practical applications in distributed, networked, cyber-physical, and autonomous systems. They are organized in topical sections on model checking, program analysis using polyhedra, synthesis, learning, runtime verification, hybrid and timed systems, tools, probabilistic systems, static analysis, theory and security, SAT, SMT and decisions procedures, concurrency, and CPS, hardware, industrial applications
    corecore