13,761 research outputs found
Complexity of the Guarded Two-Variable Fragment with Counting Quantifiers
We show that the finite satisfiability problem for the guarded two-variable
fragment with counting quantifiers is in EXPTIME. The method employed also
yields a simple proof of a result recently obtained by Y. Kazakov, that the
satisfiability problem for the guarded two-variable fragment with counting
quantifiers is in EXPTIME.Comment: 20 pages, 3 figure
Variants of Constrained Longest Common Subsequence
In this work, we consider a variant of the classical Longest Common
Subsequence problem called Doubly-Constrained Longest Common Subsequence
(DC-LCS). Given two strings s1 and s2 over an alphabet A, a set C_s of strings,
and a function Co from A to N, the DC-LCS problem consists in finding the
longest subsequence s of s1 and s2 such that s is a supersequence of all the
strings in Cs and such that the number of occurrences in s of each symbol a in
A is upper bounded by Co(a). The DC-LCS problem provides a clear mathematical
formulation of a sequence comparison problem in Computational Biology and
generalizes two other constrained variants of the LCS problem: the Constrained
LCS and the Repetition-Free LCS. We present two results for the DC-LCS problem.
First, we illustrate a fixed-parameter algorithm where the parameter is the
length of the solution. Secondly, we prove a parameterized hardness result for
the Constrained LCS problem when the parameter is the number of the constraint
strings and the size of the alphabet A. This hardness result also implies the
parameterized hardness of the DC-LCS problem (with the same parameters) and its
NP-hardness when the size of the alphabet is constant
Smoothed Complexity Theory
Smoothed analysis is a new way of analyzing algorithms introduced by Spielman
and Teng (J. ACM, 2004). Classical methods like worst-case or average-case
analysis have accompanying complexity classes, like P and AvgP, respectively.
While worst-case or average-case analysis give us a means to talk about the
running time of a particular algorithm, complexity classes allows us to talk
about the inherent difficulty of problems.
Smoothed analysis is a hybrid of worst-case and average-case analysis and
compensates some of their drawbacks. Despite its success for the analysis of
single algorithms and problems, there is no embedding of smoothed analysis into
computational complexity theory, which is necessary to classify problems
according to their intrinsic difficulty.
We propose a framework for smoothed complexity theory, define the relevant
classes, and prove some first hardness results (of bounded halting and tiling)
and tractability results (binary optimization problems, graph coloring,
satisfiability). Furthermore, we discuss extensions and shortcomings of our
model and relate it to semi-random models.Comment: to be presented at MFCS 201
Submodularity and Optimality of Fusion Rules in Balanced Binary Relay Trees
We study the distributed detection problem in a balanced binary relay tree,
where the leaves of the tree are sensors generating binary messages. The root
of the tree is a fusion center that makes the overall decision. Every other
node in the tree is a fusion node that fuses two binary messages from its child
nodes into a new binary message and sends it to the parent node at the next
level. We assume that the fusion nodes at the same level use the same fusion
rule. We call a string of fusion rules used at different levels a fusion
strategy. We consider the problem of finding a fusion strategy that maximizes
the reduction in the total error probability between the sensors and the fusion
center. We formulate this problem as a deterministic dynamic program and
express the solution in terms of Bellman's equations. We introduce the notion
of stringsubmodularity and show that the reduction in the total error
probability is a stringsubmodular function. Consequentially, we show that the
greedy strategy, which only maximizes the level-wise reduction in the total
error probability, is within a factor of the optimal strategy in terms of
reduction in the total error probability
Metamodel Instance Generation: A systematic literature review
Modelling and thus metamodelling have become increasingly important in
Software Engineering through the use of Model Driven Engineering. In this paper
we present a systematic literature review of instance generation techniques for
metamodels, i.e. the process of automatically generating models from a given
metamodel. We start by presenting a set of research questions that our review
is intended to answer. We then identify the main topics that are related to
metamodel instance generation techniques, and use these to initiate our
literature search. This search resulted in the identification of 34 key papers
in the area, and each of these is reviewed here and discussed in detail. The
outcome is that we are able to identify a knowledge gap in this field, and we
offer suggestions as to some potential directions for future research.Comment: 25 page
CHR Grammars
A grammar formalism based upon CHR is proposed analogously to the way
Definite Clause Grammars are defined and implemented on top of Prolog. These
grammars execute as robust bottom-up parsers with an inherent treatment of
ambiguity and a high flexibility to model various linguistic phenomena. The
formalism extends previous logic programming based grammars with a form of
context-sensitive rules and the possibility to include extra-grammatical
hypotheses in both head and body of grammar rules. Among the applications are
straightforward implementations of Assumption Grammars and abduction under
integrity constraints for language analysis. CHR grammars appear as a powerful
tool for specification and implementation of language processors and may be
proposed as a new standard for bottom-up grammars in logic programming.
To appear in Theory and Practice of Logic Programming (TPLP), 2005Comment: 36 pp. To appear in TPLP, 200
A new approach for diagnosability analysis of Petri nets using Verifier Nets
In this paper, we analyze the diagnosability properties of labeled Petri nets. We consider the standard notion of diagnosability of languages, requiring that every occurrence of an unobservable fault event be eventually detected, as well as the stronger notion of diagnosability in K steps, where the detection must occur within a fixed bound of K event occurrences after the fault. We give necessary and sufficient conditions for these two notions of diagnosability for both bounded and unbounded Petri nets and then present an algorithmic technique for testing the conditions based on linear programming. Our approach is novel and based on the analysis of the reachability/coverability graph of a special Petri net, called Verifier Net, that is built from the Petri net model of the given system. In the case of systems that are diagnosable in K steps, we give a procedure to compute the bound K. To the best of our knowledge, this is the first time that necessary and sufficient conditions for diagnosability and diagnosability in K steps of labeled unbounded Petri nets are presented
An optimal quantum algorithm for the oracle identification problem
In the oracle identification problem, we are given oracle access to an
unknown N-bit string x promised to belong to a known set C of size M and our
task is to identify x. We present a quantum algorithm for the problem that is
optimal in its dependence on N and M. Our algorithm considerably simplifies and
improves the previous best algorithm due to Ambainis et al. Our algorithm also
has applications in quantum learning theory, where it improves the complexity
of exact learning with membership queries, resolving a conjecture of Hunziker
et al.
The algorithm is based on ideas from classical learning theory and a new
composition theorem for solutions of the filtered -norm semidefinite
program, which characterizes quantum query complexity. Our composition theorem
is quite general and allows us to compose quantum algorithms with
input-dependent query complexities without incurring a logarithmic overhead for
error reduction. As an application of the composition theorem, we remove all
log factors from the best known quantum algorithm for Boolean matrix
multiplication.Comment: 16 pages; v2: minor change
- …