140,807 research outputs found
Clean Clothes: International Meeting Barcelona 2001
CCC_intlmeetingSKO_bracelonadef.pdf: 104 downloads, before Oct. 1, 2020
Clean Clothes â International Meeting: Barcelona 2001
Brochure reporting on the evaluation the CCC conducted of its âaims and activities, their impact, and the way in which the campaign functions and is organized.â The report is divided into five parts: evaluation, strategic concerns, future trends, regional perspectives, and an agenda for action
Domain-Type-Guided Refinement Selection Based on Sliced Path Prefixes
Abstraction is a successful technique in software verification, and
interpolation on infeasible error paths is a successful approach to
automatically detect the right level of abstraction in counterexample-guided
abstraction refinement. Because the interpolants have a significant influence
on the quality of the abstraction, and thus, the effectiveness of the
verification, an algorithm for deriving the best possible interpolants is
desirable. We present an analysis-independent technique that makes it possible
to extract several alternative sequences of interpolants from one given
infeasible error path, if there are several reasons for infeasibility in the
error path. We take as input the given infeasible error path and apply a
slicing technique to obtain a set of error paths that are more abstract than
the original error path but still infeasible, each for a different reason. The
(more abstract) constraints of the new paths can be passed to a standard
interpolation engine, in order to obtain a set of interpolant sequences, one
for each new path. The analysis can then choose from this set of interpolant
sequences and select the most appropriate, instead of being bound to the single
interpolant sequence that the interpolation engine would normally return. For
example, we can select based on domain types of variables in the interpolants,
prefer to avoid loop counters, or compare with templates for potential loop
invariants, and thus control what kind of information occurs in the abstraction
of the program. We implemented the new algorithm in the open-source
verification framework CPAchecker and show that our proof-technique-independent
approach yields a significant improvement of the effectiveness and efficiency
of the verification process.Comment: 10 pages, 5 figures, 1 table, 4 algorithm
Shaded Tangles for the Design and Verification of Quantum Programs (Extended Abstract)
We give a scheme for interpreting shaded tangles as quantum programs, with
the property that isotopic tangles yield equivalent programs. We analyze many
known quantum programs in this way -- including entanglement manipulation and
error correction -- and in each case present a fully-topological formal
verification, yielding in several cases substantial new insight into how the
program works. We also use our methods to identify several new or generalized
procedures.Comment: In Proceedings QPL 2017, arXiv:1802.0973
Recommended from our members
M&V in ESPC: The U.S. Federal Experience and Implications for Developing ESPC Markets
The United States Federal Government has been conducting guaranteed savings energy savings performance contracts for over 20 years and now relies on ESPC for the majority of its energy efficiency work. Along with a related financed project type, these deals resulted in $4.2 billion of project investment in the five years ending in 2016, a pace that has even accelerated since.
Measurement and verification (M&V) on the projects is the key to assuring savings realization and persistence. Perceived as a weakness or burdensome added cost in the early years of the program, M&V has become a strength. All energy conservation measures (ECMs) have some form of measurement â defined as a measured baseline establishment followed by at least one measurement of the main energy-saving parameter taken in the performance period for each ECM. The governmentâs in-house energy consulting office, the Federal Energy Management Program (FEMP), now recommends measurement of these âOption Aâ M&V ECMs throughout the contract term, usually annually. Moreover, a significantly higher percentage of projects are now characterized by more ambitious M&V, including Option B (all parameter measurement) for most generation (including renewable) and some efficiency measures, and more frequent Option C (whole facility utility bill analysis) for âdeep retrofitâ projects with multiple, interactive ECMs. Coincident with this progress in M&V has been a much greater embracing of ESPC by the federal agencies, resulting in the enormous rate of projects now executed.
This paper traces the evolution of M&V in federal ESPC and argues that the heightened credibility of the savings has contributed significantly to the procurement vehicleâs long-term viability. This focus on savings integrity via M&V has been learned over two decades for U.S. federal ESPC, but countries with developing ESPC markets would be wise to emphasize it as their markets emerge, allowing them to avoid some of the âgrowing painsâ experienced in the U.S
Quantitative Safety: Linking Proof-Based Verification with Model Checking for Probabilistic Systems
This paper presents a novel approach for augmenting proof-based verification
with performance-style analysis of the kind employed in state-of-the-art model
checking tools for probabilistic systems. Quantitative safety properties
usually specified as probabilistic system invariants and modeled in proof-based
environments are evaluated using bounded model checking techniques.
Our specific contributions include the statement of a theorem that is central
to model checking safety properties of proof-based systems, the establishment
of a procedure; and its full implementation in a prototype system (YAGA) which
readily transforms a probabilistic model specified in a proof-based environment
to its equivalent verifiable PRISM model equipped with reward structures. The
reward structures capture the exact interpretation of the probabilistic
invariants and can reveal succinct information about the model during
experimental investigations. Finally, we demonstrate the novelty of the
technique on a probabilistic library case study
- âŠ