157 research outputs found
Verifying safety properties of a nonlinear control by interactive theorem proving with the Prototype Verification System
Interactive, or computer-assisted, theorem proving is the verification of statements in a
formal system, where the proof is developed by a logician who chooses the appropriate inference steps, in turn executed by an automatic theorem prover. In this paper, interactive theorem proving is used to verify safety properties of a nonlinear (hybrid) control system
Helac-Phegas: a generator for all parton level processes
The updated version of the Helac-Phegas event generator is presented. The
matrix elements are calculated through Dyson-Schwinger recursive equations.
Helac-Phegas generates parton-level events with all necessary information, in
the most recent Les Houches Accord format, for the study of any process within
the Standard Model in hadron and lepton colliders
Recommended from our members
Message Queue Manager
The conventional way of debugging is to examine the program's state at various points during execution. With the addition of message passing as a means of communication and synchronization in message passing programs, the state of pending message operations also needs to be examined. Current parallel debuggers or analysis tools offer little information on the status of message operations. When they do, the information presented is voluminous and hard to assimilate.
The Message Queue Manager (MQM), a Parallel Tools Consortium project, provides a simple and direct way to determine the status of pending message operations during program execution. It has been designed and developed as a modular component that can be incorporated into a parallel debugger or other analysis tools on a variety of hardware platforms. MQM obtains a snapshot of message queues from the underlying message passing system (using a standard interface at some breakpoint or other pause in execution) and presents the data graphically. Two levels of information are available -- an overview gives an abstraction of message queue lengths, while a detailed view shows pending message operations for specific processes. The overview can display a large number of processes in a reasonably small screen, accommodating massively parallel programs. Several filtering mechanisms allow the user easily control how much information is shown
A Migratable User-Level Process Package for PVM
Shared, multi-user, workstation networks are characterized by unpredictable variability in system load. Further, the concept of workstation ownership is typically present. For efficient and unobtrusive computing in such environments, applications must not only overlap their computation with communication but also redistribute their computations adaptively based on changes in workstation availability and load. Managing these issues at application level leads to programs that are difficult to write and debug. In this paper, we present a system that manages this dynamic multi-processor environment while exporting a simple message-based programming model of a dedicated, distributed memory multiprocessor to applications. Programmers are thus insulated from the many complexities of the dynamic environment at the same time are able to achieve the benefits of multi-threading, adaptive load distribution and unobtrusive computing. To support the dedicated multi-processor model efficiently, the system defines a new kind of virtual processor called User-Level Process (ULP) that can be used to implement efficient multi-threading and application-transparent migration. The viability of ULPs is demonstrated through UPVM, a prototype implementation of the PVM message passing interface using ULPs. Typically, existing PVM programs written in Single Program Multiple Data (SPMD) style need only be re-compiled to use this package. The design of the package is presented and the performance analyzed with respect to both micro-benchmarks and some complete PVM applications. Finally, we discuss aspects of the ULP package that affect its portability and its support for heterogeneity, application transparency, and application debugging
A gentle transition from Java programming to Web Services using XML-RPC
Exposing students to leading edge vocational areas of relevance such as Web Services can be difficult. We show a lightweight approach by embedding a key component of Web Services within a Level 3 BSc module in Distributed Computing. We present a ready to use collection of lecture slides and student activities based on XML-RPC. In
addition we show that this material addresses the central topics in the context of web services as identified by Draganova (2003)
Markovian Monte Carlo program EvolFMC v.2 for solving QCD evolution equations
We present the program EvolFMC v.2 that solves the evolution equations in QCD
for the parton momentum distributions by means of the Monte Carlo technique
based on the Markovian process. The program solves the DGLAP-type evolution as
well as modified-DGLAP ones. In both cases the evolution can be performed in
the LO or NLO approximation. The quarks are treated as massless. The overall
technical precision of the code has been established at 0.05% precision level.
This way, for the first time ever, we demonstrate that with the Monte Carlo
method one can solve the evolution equations with precision comparable to the
other numerical methods.Comment: 38 pages, 9 Postscript figure
Workload based provenance capture reduction
Multiple solutions have been developed that collect provenance in Data-Intensive Scalable Computing (DISC) systems like Apache Spark and Apache Hadoop. Existing solutions include RAMP, Newt, Lipstick and Titian. Though these solutions support debugging within the dataflow programs, they introduce a space overhead of 30-50% of the size of the input data during provenance collection. In a productive environment, this overhead is too high to permanently track provenance and to store all the provenance information. That is why solutions exist that reduce the amount of provenance data after their collection. Among those are Prox, Propolis and distillations. However, they do not address the problem of incurring space overhead during the execution of a dataflow program. The existing provenance reduction techniques do not consider optimizing the provenance reduction based on particular use cases or applications of provenance. The goal of this thesis is to find and evaluate application-dependent provenance data reduction techniques that are applicable during execution of dataflow programs. To this end, we survey multiple applications and use cases of provenance like data exploration, monitoring, data quality etc. In addition, we analyze how provenance is being used in them. Furthermore, we introduce nine data reduction techniques that can be applied to provenance in the context of different use cases.
We formally describe and evaluate four out of the nine techniques - sampling, histogram, clustering and equivalence classes on top of Apache Spark. There is no benchmark available to test different provenance solutions. Hence, we define six scenarios on two different datasets to evaluate them. We also consider the application of provenance in each scenario. We use these techniques to obtain reduced provenance data then, we introduce three metrics to compare the reduced provenance data to full provenance. We perform a quantitative analysis to compare different techniques based on these metrics. Afterwards, we perform a qualitative analysis to examine the effectiveness of different reduction techniques in the context of a particular use case
Monitoring Teams by Overhearing: A Multi-Agent Plan-Recognition Approach
Recent years are seeing an increasing need for on-line monitoring of teams of
cooperating agents, e.g., for visualization, or performance tracking. However,
in monitoring deployed teams, we often cannot rely on the agents to always
communicate their state to the monitoring system. This paper presents a
non-intrusive approach to monitoring by 'overhearing', where the monitored
team's state is inferred (via plan-recognition) from team-members' routine
communications, exchanged as part of their coordinated task execution, and
observed (overheard) by the monitoring system. Key challenges in this approach
include the demanding run-time requirements of monitoring, the scarceness of
observations (increasing monitoring uncertainty), and the need to scale-up
monitoring to address potentially large teams. To address these, we present a
set of complementary novel techniques, exploiting knowledge of the social
structures and procedures in the monitored team: (i) an efficient probabilistic
plan-recognition algorithm, well-suited for processing communications as
observations; (ii) an approach to exploiting knowledge of the team's social
behavior to predict future observations during execution (reducing monitoring
uncertainty); and (iii) monitoring algorithms that trade expressivity for
scalability, representing only certain useful monitoring hypotheses, but
allowing for any number of agents and their different activities to be
represented in a single coherent entity. We present an empirical evaluation of
these techniques, in combination and apart, in monitoring a deployed team of
agents, running on machines physically distributed across the country, and
engaged in complex, dynamic task execution. We also compare the performance of
these techniques to human expert and novice monitors, and show that the
techniques presented are capable of monitoring at human-expert levels, despite
the difficulty of the task
MACE: Joint Deliverable "Evaluation of the MACE system"
Stefaner, M., Wolpers, M., Memmel, M., Duval, E., Specht, M., Börner, D., Gruber, M., De Jong, T., Giretti, A., & Klemke, R. (2009). MACE: Joint Deliverable "Evaluation of the MACE system". MACE-project.This deliverable presents the evaluation results of the MACE project1. MACE integrates large amounts of technology enhanced learning contents and metadata in the domain of architecture, in order to create a framework for the integration of multiple content sources, content enrichment with different metadata types, and improved access to these existing contents.MAC
- …