2,202 research outputs found
Generalizing Consistency and other Constraint Properties to Quantified Constraints
Quantified constraints and Quantified Boolean Formulae are typically much
more difficult to reason with than classical constraints, because quantifier
alternation makes the usual notion of solution inappropriate. As a consequence,
basic properties of Constraint Satisfaction Problems (CSP), such as consistency
or substitutability, are not completely understood in the quantified case.
These properties are important because they are the basis of most of the
reasoning methods used to solve classical (existentially quantified)
constraints, and one would like to benefit from similar reasoning methods in
the resolution of quantified constraints. In this paper, we show that most of
the properties that are used by solvers for CSP can be generalized to
quantified CSP. This requires a re-thinking of a number of basic concepts; in
particular, we propose a notion of outcome that generalizes the classical
notion of solution and on which all definitions are based. We propose a
systematic study of the relations which hold between these properties, as well
as complexity results regarding the decision of these properties. Finally, and
since these problems are typically intractable, we generalize the approach used
in CSP and propose weaker, easier to check notions based on locality, which
allow to detect these properties incompletely but in polynomial time
Certainty Closure: Reliable Constraint Reasoning with Incomplete or Erroneous Data
Constraint Programming (CP) has proved an effective paradigm to model and
solve difficult combinatorial satisfaction and optimisation problems from
disparate domains. Many such problems arising from the commercial world are
permeated by data uncertainty. Existing CP approaches that accommodate
uncertainty are less suited to uncertainty arising due to incomplete and
erroneous data, because they do not build reliable models and solutions
guaranteed to address the user's genuine problem as she perceives it. Other
fields such as reliable computation offer combinations of models and associated
methods to handle these types of uncertain data, but lack an expressive
framework characterising the resolution methodology independently of the model.
We present a unifying framework that extends the CP formalism in both model
and solutions, to tackle ill-defined combinatorial problems with incomplete or
erroneous data. The certainty closure framework brings together modelling and
solving methodologies from different fields into the CP paradigm to provide
reliable and efficient approches for uncertain constraint problems. We
demonstrate the applicability of the framework on a case study in network
diagnosis. We define resolution forms that give generic templates, and their
associated operational semantics, to derive practical solution methods for
reliable solutions.Comment: Revised versio
Verifying UML/OCL operation contracts
In current model-driven development approaches, software models are the primary artifacts of the development process. Therefore, assessment of their correctness is a key issue to ensure the quality of the final application. Research on model consistency has focused mostly on the models' static aspects. Instead, this paper addresses the verification of their dynamic aspects, expressed as a set of operations defined by means of pre/postcondition contracts. This paper presents an automatic method based on Constraint Programming to verify UML models extended with OCL constraints and operation contracts. In our approach, both static and dynamic aspects are translated into a Constraint Satisfaction Problem. Then, compliance of the operations with respect to several correctness properties such as operation executability or determinism are formally verified
Set-based design of mechanical systems with design robustness integrated
This paper presents a method for parameter design of mechanical products based on a set-based approach. Set-based concurrent engineering emphasises on designing in a multi-stakeholder environment with concurrent involvement of the stakeholders in the design process. It also encourages flexibility in design through communication in terms of ranges instead of fixed point values and subsequent alternative solutions resulting from intersection of these ranges. These alternative solutions can then be refined and selected according to the designers’ preferences and clients’ needs. This paper presents a model and tools for integrated flexible design that take into account the manufacturing variations as well as the design objectives for finding inherently robust solutions using QCSP transformation through interval analysis. In order to demonstrate the approach, an example of design of rigid flange coupling with a variable number of bolts and a choice of bolts from ISO M standard has been resolved and demonstrated
Linear Datalog and Bounded Path Duality of Relational Structures
In this paper we systematically investigate the connections between logics
with a finite number of variables, structures of bounded pathwidth, and linear
Datalog Programs. We prove that, in the context of Constraint Satisfaction
Problems, all these concepts correspond to different mathematical embodiments
of a unique robust notion that we call bounded path duality. We also study the
computational complexity implications of the notion of bounded path duality. We
show that every constraint satisfaction problem \csp(\best) with bounded path
duality is solvable in NL and that this notion explains in a uniform way all
families of CSPs known to be in NL. Finally, we use the results developed in
the paper to identify new problems in NL
On the Scope of the Universal-Algebraic Approach to Constraint Satisfaction
The universal-algebraic approach has proved a powerful tool in the study of
the complexity of CSPs. This approach has previously been applied to the study
of CSPs with finite or (infinite) omega-categorical templates, and relies on
two facts. The first is that in finite or omega-categorical structures A, a
relation is primitive positive definable if and only if it is preserved by the
polymorphisms of A. The second is that every finite or omega-categorical
structure is homomorphically equivalent to a core structure. In this paper, we
present generalizations of these facts to infinite structures that are not
necessarily omega-categorical. (This abstract has been severely curtailed by
the space constraints of arXiv -- please read the full abstract in the
article.) Finally, we present applications of our general results to the
description and analysis of the complexity of CSPs. In particular, we give
general hardness criteria based on the absence of polymorphisms that depend on
more than one argument, and we present a polymorphism-based description of
those CSPs that are first-order definable (and therefore can be solved in
polynomial time).Comment: Extended abstract appeared at 25th Symposium on Logic in Computer
Science (LICS 2010). This version will appear in the LMCS special issue
associated with LICS 201
Lazy Model Expansion: Interleaving Grounding with Search
Finding satisfying assignments for the variables involved in a set of
constraints can be cast as a (bounded) model generation problem: search for
(bounded) models of a theory in some logic. The state-of-the-art approach for
bounded model generation for rich knowledge representation languages, like ASP,
FO(.) and Zinc, is ground-and-solve: reduce the theory to a ground or
propositional one and apply a search algorithm to the resulting theory.
An important bottleneck is the blowup of the size of the theory caused by the
reduction phase. Lazily grounding the theory during search is a way to overcome
this bottleneck. We present a theoretical framework and an implementation in
the context of the FO(.) knowledge representation language. Instead of
grounding all parts of a theory, justifications are derived for some parts of
it. Given a partial assignment for the grounded part of the theory and valid
justifications for the formulas of the non-grounded part, the justifications
provide a recipe to construct a complete assignment that satisfies the
non-grounded part. When a justification for a particular formula becomes
invalid during search, a new one is derived; if that fails, the formula is
split in a part to be grounded and a part that can be justified.
The theoretical framework captures existing approaches for tackling the
grounding bottleneck such as lazy clause generation and grounding-on-the-fly,
and presents a generalization of the 2-watched literal scheme. We present an
algorithm for lazy model expansion and integrate it in a model generator for
FO(ID), a language extending first-order logic with inductive definitions. The
algorithm is implemented as part of the state-of-the-art FO(ID) Knowledge-Base
System IDP. Experimental results illustrate the power and generality of the
approach
Quantum walk speedup of backtracking algorithms
We describe a general method to obtain quantum speedups of classical
algorithms which are based on the technique of backtracking, a standard
approach for solving constraint satisfaction problems (CSPs). Backtracking
algorithms explore a tree whose vertices are partial solutions to a CSP in an
attempt to find a complete solution. Assume there is a classical backtracking
algorithm which finds a solution to a CSP on n variables, or outputs that none
exists, and whose corresponding tree contains T vertices, each vertex
corresponding to a test of a partial solution. Then we show that there is a
bounded-error quantum algorithm which completes the same task using O(sqrt(T)
n^(3/2) log n) tests. In particular, this quantum algorithm can be used to
speed up the DPLL algorithm, which is the basis of many of the most efficient
SAT solvers used in practice. The quantum algorithm is based on the use of a
quantum walk algorithm of Belovs to search in the backtracking tree. We also
discuss how, for certain distributions on the inputs, the algorithm can lead to
an exponential reduction in expected runtime.Comment: 23 pages; v2: minor changes to presentatio
A Graph-Based Semantics Workbench for Concurrent Asynchronous Programs
A number of novel programming languages and libraries have been proposed that
offer simpler-to-use models of concurrency than threads. It is challenging,
however, to devise execution models that successfully realise their
abstractions without forfeiting performance or introducing unintended
behaviours. This is exemplified by SCOOP---a concurrent object-oriented
message-passing language---which has seen multiple semantics proposed and
implemented over its evolution. We propose a "semantics workbench" with fully
and semi-automatic tools for SCOOP, that can be used to analyse and compare
programs with respect to different execution models. We demonstrate its use in
checking the consistency of semantics by applying it to a set of representative
programs, and highlighting a deadlock-related discrepancy between the principal
execution models of the language. Our workbench is based on a modular and
parameterisable graph transformation semantics implemented in the GROOVE tool.
We discuss how graph transformations are leveraged to atomically model
intricate language abstractions, and how the visual yet algebraic nature of the
model can be used to ascertain soundness.Comment: Accepted for publication in the proceedings of FASE 2016 (to appear
- …