511 research outputs found
A Lower Bound for Sampling Disjoint Sets
Suppose Alice and Bob each start with private randomness and no other input, and they wish to engage in a protocol in which Alice ends up with a set x subseteq[n] and Bob ends up with a set y subseteq[n], such that (x,y) is uniformly distributed over all pairs of disjoint sets. We prove that for some constant beta0 of the uniform distribution over all pairs of disjoint sets of size sqrt{n}
A Robust Class of Linear Recurrence Sequences
We introduce a subclass of linear recurrence sequences which we call poly-rational sequences because they are denoted by rational expressions closed under sum and product. We show that this class is robust by giving several characterisations: polynomially ambiguous weighted automata, copyless cost-register automata, rational formal series, and linear recurrence sequences whose eigenvalues are roots of rational numbers
Unambiguous Separators for Tropical Tree Automata
In this paper we show that given a max-plus automaton (over trees, and with real weights) computing a function f and a min-plus automaton (similar) computing a function g such that f ? g, there exists effectively an unambiguous tropical automaton computing h such that f ? h ? g.
This generalizes a result of Lombardy and Mairesse of 2006 stating that series which are both max-plus and min-plus rational are unambiguous. This generalization goes in two directions: trees are considered instead of words, and separation is established instead of characterization (separation implies characterization). The techniques in the two proofs are very different
Lower Bound Techniques for QBF Proof Systems
How do we prove that a false QBF is inded false? How big a proof is needed? The special case when all quantifiers are existential is the well-studied setting of propositional proof complexity. Expectedly, universal quantifiers change the game significantly. Several proof systems have been designed in the last couple of decades to handle QBFs. Lower bound paradigms from propositional proof complexity cannot always be extended - in most cases feasible interpolation and consequent transfer of circuit lower bounds works, but obtaining lower bounds on size by providing lower bounds on width fails dramatically. A new paradigm with no analogue in the propositional world has emerged in the form of strategy extraction, allowing for transfer of circuit lower bounds, as well as obtaining independent
genuine QBF lower bounds based on a semantic cost measure.
This talk will provide a broad overview of some of these developments
Low Randomness Rumor Spreading via Hashing
International audienceWe consider the classical rumor spreading problem, where a piece of information must be disseminated from a single node to all n nodes of a given network. We devise two simple push-based protocols, in which nodes choose the neighbor they send the information to in each round using pairwise independent hash functions, or a pseudo-random generator, respectively. For several well-studied topologies our algorithms use exponentially fewer random bits than previous protocols. For example, in complete graphs, expanders, and random graphs only a polylogarithmic number of random bits are needed in total to spread the rumor in O(log n) rounds with high probability. Previous explicit algorithms require Omega(n) random bits to achieve the same round complexity. For complete graphs, the amount of randomness used by our hashing-based algorithm is within an O(log n)-factor of the theoretical minimum determined by [Giakkoupis and Woelfel, 2011]
Counting edge-injective homomorphisms and matchings on restricted graph classes
We consider the -hard problem of counting all matchings with
exactly edges in a given input graph ; we prove that it remains
-hard on graphs that are line graphs or bipartite graphs
with degree on one side. In our proofs, we use that -matchings in line
graphs can be equivalently viewed as edge-injective homomorphisms from the
disjoint union of length- paths into (arbitrary) host graphs. Here, a
homomorphism from to is edge-injective if it maps any two distinct
edges of to distinct edges in . We show that edge-injective
homomorphisms from a pattern graph can be counted in polynomial time if
has bounded vertex-cover number after removing isolated edges. For hereditary
classes of pattern graphs, we complement this result: If the
graphs in have unbounded vertex-cover number even after deleting
isolated edges, then counting edge-injective homomorphisms with patterns from
is -hard. Our proofs rely on an edge-colored
variant of Holant problems and a delicate interpolation argument; both may be
of independent interest.Comment: 35 pages, 9 figure
Tree Automata with Global Constraints for Infinite Trees
We study an extension of tree automata on infinite trees with global equality and disequality constraints. These constraints can enforce that all subtrees for which in the accepting run a state q is reached (at the root of that subtree) are identical, or that these trees differ from the subtrees at which a state q\u27 is reached. We consider the closure properties of this model and its decision problems. While the emptiness problem for the general model remains open, we show the decidability of the emptiness problem for the case that the given automaton only uses equality constraints
The Tandem Duplication Distance Is NP-Hard
In computational biology, tandem duplication is an important biological phenomenon which can occur either at the genome or at the DNA level. A tandem duplication takes a copy of a genome segment and inserts it right after the segment - this can be represented as the string operation AXB ? AXXB. Tandem exon duplications have been found in many species such as human, fly or worm, and have been largely studied in computational biology.
The Tandem Duplication (TD) distance problem we investigate in this paper is defined as follows: given two strings S and T over the same alphabet, compute the smallest sequence of tandem duplications required to convert S to T. The natural question of whether the TD distance can be computed in polynomial time was posed in 2004 by Leupold et al. and had remained open, despite the fact that tandem duplications have received much attention ever since. In this paper, we prove that this problem is NP-hard, settling the 16-year old open problem. We further show that this hardness holds even if all characters of S are distinct. This is known as the exemplar TD distance, which is of special relevance in bioinformatics. One of the tools we develop for the reduction is a new problem called the Cost-Effective Subgraph, for which we obtain W[1]-hardness results that might be of independent interest. We finally show that computing the exemplar TD distance between S and T is fixed-parameter tractable. Our results open the door to many other questions, and we conclude with several open problems
- âŠ