2,997 research outputs found

    Automating synthesis of asynchronous communication mechanisms

    Get PDF
    Asynchronous data communication mechanisms (ACMs) have been extensively studied as data connectors between independently timed processes in digital systems. In previous work, systematic ACM synthesis methods have been proposed. In this paper, we advance this work by developing algorithms and software tools which automate the major part of the ACM synthesis process. Firstly, an interleaving specification is constructed in the form of a state graph, and secondly, a Petri net model of an "ACM-type" is derived using the notion of an ACM-region. The method is applied to a number of "standard" writing and reading policies of ACMs with shared memory and unidirectional control variables.Peer ReviewedPostprint (published version

    Special Session on Industry 4.0

    Get PDF
    No abstract available

    Distributed, cooperating knowledge-based systems

    Get PDF
    Some current research in the development and application of distributed, cooperating knowledge-based systems technology is addressed. The focus of the current research is the spacecraft ground operations environment. The underlying hypothesis is that, because of the increasing size, complexity, and cost of planned systems, conventional procedural approaches to the architecture of automated systems will give way to a more comprehensive knowledge-based approach. A hallmark of these future systems will be the integration of multiple knowledge-based agents which understand the operational goals of the system and cooperate with each other and the humans in the loop to attain the goals. The current work includes the development of a reference model for knowledge-base management, the development of a formal model of cooperating knowledge-based agents, the use of testbed for prototyping and evaluating various knowledge-based concepts, and beginning work on the establishment of an object-oriented model of an intelligent end-to-end (spacecraft to user) system. An introductory discussion of these activities is presented, the major concepts and principles being investigated are highlighted, and their potential use in other application domains is indicated

    On Synchronous and Asynchronous Monitor Instrumentation for Actor-based systems

    Full text link
    We study the impact of synchronous and asynchronous monitoring instrumentation on runtime overheads in the context of a runtime verification framework for actor-based systems. We show that, in such a context, asynchronous monitoring incurs substantially lower overhead costs. We also show how, for certain properties that require synchronous monitoring, a hybrid approach can be used that ensures timely violation detections for the important events while, at the same time, incurring lower overhead costs that are closer to those of an asynchronous instrumentation.Comment: In Proceedings FOCLASA 2014, arXiv:1502.0315

    Coupling as a trade-off in an Enterprise Service Bus

    Get PDF
    Traditionally, integration problems between IT systems were solved by point-to-point connections. These point-to-point connections pose issues with scalability, reliability, and flexibility. To overcome these issues, companies typically invest in Enterprise Application Integration (EAI) using an Enterprise Service Bus (ESB) to integrate the IT systems through a central middleware infrastructure. EAI promises improvement of scalability, reliability, and flexibility by implementing loosely coupled integration solutions to realise loosely coupled IT systems. By wrongly implementing EAI on an ESB IT systems may still be tightly coupled and the issues with point-to-point connections could be recreated on the ESB. Currently there is no out-of-the-box solution to identify the integration solution where tight coupling causes these issues. The goal of this research is to investigate an approach to identify the coupling state in an Enterprise Service Bus and identify the integration solutions on an ESB which have a negative impact on the quality attributes due to tight coupling. The first step in the approach is applying a set of properties on the integration solutions to identify their coupling state. Manually identifying the coupling state is labour intensive, so it is automated by implementing a prototype with the Eclipse MoDisco framework. The second step in the approach is evaluating a trade-off between the risk of being in a certain coupling state and the efficiency loss of migrating to a less risky coupling state. With the outcome of the trade-off it can be ascertained whether or not it is beneficial to migrate to a different coupling state. The result of the approach is a list of integration solutions for which it would be beneficial to migrate to a different coupling state. This gives a concrete measure to be able to determine which integration solutions need to be improved to strive for the optimal balance between quality and the effort needed to realise quality. The approach was validated using the ESB implementation of a large European airport as a case study

    Synthesis of behavioral models from scenarios

    No full text

    Adaptive Process Management in Cyber-Physical Domains

    Get PDF
    The increasing application of process-oriented approaches in new challenging cyber-physical domains beyond business computing (e.g., personalized healthcare, emergency management, factories of the future, home automation, etc.) has led to reconsider the level of flexibility and support required to manage complex processes in such domains. A cyber-physical domain is characterized by the presence of a cyber-physical system coordinating heterogeneous ICT components (PCs, smartphones, sensors, actuators) and involving real world entities (humans, machines, agents, robots, etc.) that perform complex tasks in the “physical” real world to achieve a common goal. The physical world, however, is not entirely predictable, and processes enacted in cyber-physical domains must be robust to unexpected conditions and adaptable to unanticipated exceptions. This demands a more flexible approach in process design and enactment, recognizing that in real-world environments it is not adequate to assume that all possible recovery activities can be predefined for dealing with the exceptions that can ensue. In this chapter, we tackle the above issue and we propose a general approach, a concrete framework and a process management system implementation, called SmartPM, for automatically adapting processes enacted in cyber-physical domains in case of unanticipated exceptions and exogenous events. The adaptation mechanism provided by SmartPM is based on declarative task specifications, execution monitoring for detecting failures and context changes at run-time, and automated planning techniques to self-repair the running process, without requiring to predefine any specific adaptation policy or exception handler at design-time

    Data Provenance and Management in Radio Astronomy: A Stream Computing Approach

    Get PDF
    New approaches for data provenance and data management (DPDM) are required for mega science projects like the Square Kilometer Array, characterized by extremely large data volume and intense data rates, therefore demanding innovative and highly efficient computational paradigms. In this context, we explore a stream-computing approach with the emphasis on the use of accelerators. In particular, we make use of a new generation of high performance stream-based parallelization middleware known as InfoSphere Streams. Its viability for managing and ensuring interoperability and integrity of signal processing data pipelines is demonstrated in radio astronomy. IBM InfoSphere Streams embraces the stream-computing paradigm. It is a shift from conventional data mining techniques (involving analysis of existing data from databases) towards real-time analytic processing. We discuss using InfoSphere Streams for effective DPDM in radio astronomy and propose a way in which InfoSphere Streams can be utilized for large antennae arrays. We present a case-study: the InfoSphere Streams implementation of an autocorrelating spectrometer, and using this example we discuss the advantages of the stream-computing approach and the utilization of hardware accelerators
    • …
    corecore