32 research outputs found
A Superpolynomial Lower Bound for the Size of Non-Deterministic Complement of an Unambiguous Automaton
Unambiguous non-deterministic finite automata (UFA) are non-deterministic automata (over finite words) such that there is at most one accepting run over each input. Such automata are known to be potentially exponentially more succinct than deterministic automata, and non-deterministic automata can be exponentially more succinct than them.
In this paper we establish a superpolynomial lower bound for the state complexity of the translation of an UFA to a non-deterministic automaton for the complement language. This disproves the formerly conjectured polynomial upper bound for this translation. This lower bound only involves a one letter alphabet, and makes use of the random graph methods.
The same proof also shows that the translation of sweeping automata to non-deterministic automata is superpolynomial
Two-Way Automata Making Choices Only at the Endmarkers
The question of the state-size cost for simulation of two-way
nondeterministic automata (2NFAs) by two-way deterministic automata (2DFAs) was
raised in 1978 and, despite many attempts, it is still open. Subsequently, the
problem was attacked by restricting the power of 2DFAs (e.g., using a
restricted input head movement) to the degree for which it was already possible
to derive some exponential gaps between the weaker model and the standard
2NFAs. Here we use an opposite approach, increasing the power of 2DFAs to the
degree for which it is still possible to obtain a subexponential conversion
from the stronger model to the standard 2DFAs. In particular, it turns out that
subexponential conversion is possible for two-way automata that make
nondeterministic choices only when the input head scans one of the input tape
endmarkers. However, there is no restriction on the input head movement. This
implies that an exponential gap between 2NFAs and 2DFAs can be obtained only
for unrestricted 2NFAs using capabilities beyond the proposed new model. As an
additional bonus, conversion into a machine for the complement of the original
language is polynomial in this model. The same holds for making such machines
self-verifying, halting, or unambiguous. Finally, any superpolynomial lower
bound for the simulation of such machines by standard 2DFAs would imply LNL.
In the same way, the alternating version of these machines is related to L =?
NL =? P, the classical computational complexity problems.Comment: 23 page
The Containment Problem for Unambiguous Register Automata
We investigate the complexity of the containment problem "Does L(A)subseteq L(B) hold?", where B is an unambiguous register automaton and A is an arbitrary register automaton. We prove that the problem is decidable and give upper bounds on the computational complexity in the general case, and when B is restricted to have a fixed number of registers
Universality Problem for Unambiguous VASS
We study languages of unambiguous VASS, that is, Vector Addition Systems with States, whose transitions read letters from a finite alphabet, and whose acceptance condition is defined by a set of final states (i.e., the coverability language). We show that the problem of universality for unambiguous VASS is ExpSpace-complete, in sheer contrast to Ackermann-completeness for arbitrary VASS, even in dimension 1. When the dimension d ? ? is fixed, the universality problem is PSpace-complete if d ? 2, and coNP-hard for 1-dimensional VASSes (also known as One Counter Nets)
Bidimensional Linear Recursive Sequences and Universality of Unambiguous Register Automata
We study the universality and inclusion problems for register automata over
equality data. We show that the universality and the inclusion problems can be
solved with 2-EXPTIME complexity when the input automata are without guessing
and unambiguous, improving on the currently best-known 2-EXPSPACE upper bound
by Mottet and Quaas. When the number of registers of both automata is fixed, we
obtain a lower EXPTIME complexity, also improving the EXPSPACE upper bound from
Mottet and Quaas for fixed number of registers. We reduce inclusion to
universality, and then we reduce universality to the problem of counting the
number of orbits of runs of the automaton. We show that the orbit-counting
function satisfies a system of bidimensional linear recursive equations with
polynomial coefficients (linrec), which generalises analogous recurrences for
the Stirling numbers of the second kind, and then we show that universality
reduces to the zeroness problem for linrec sequences. While such a counting
approach is classical and has successfully been applied to unambiguous finite
automata and grammars over finite alphabets, its application to register
automata over infinite alphabets is novel. We provide two algorithms to decide
the zeroness problem for bidimensional linear recursive sequences arising from
orbit-counting functions. Both algorithms rely on techniques from linear
non-commutative algebra. The first algorithm performs variable elimination and
has elementary complexity. The second algorithm is a refined version of the
first one and it relies on the computation of the Hermite normal form of
matrices over a skew polynomial field. The second algorithm yields an EXPTIME
decision procedure for the zeroness problem of linrec sequences, which in turn
yields the claimed bounds for the universality and inclusion problems of
register automata.Comment: full version of the homonymous paper to appear in the proceedings of
STACS'2
DESCRIPTIONAL COMPLEXITY AND PARIKH EQUIVALENCE
The thesis deals with some topics in the theory of formal languages and automata. Speci\ufb01cally, the thesis deals with the theory of context-free languages and the study of their descriptional complexity.
The descriptional complexity of a formal structure (e.g., grammar, model of automata, etc) is the number of symbols needed to write down its description. While this aspect is extensively treated in regular languages, as evidenced by numerous references, in the case of context-free languages few results are known.
An important result in this area is the Parikh\u2019s theorem. The theorem states that for each context-free language there exists a regular language with the same Parikh image. Given an alphabet \u3a3 = {a1, . . . , am}, the Parikh image is a function \u3c8 : \u3a3^ 17\u2192 N^m that associates with each word w 08\u3a3^ 17, the vector \u3c8(w)=(|w|_a1, |w|_a2, . . . , |w|_am), where |w|_ai is the number of occurrences of ai in w. The Parikh image of a language L 86\u3a3^ 17 is the set of Parikh images of its words. For instance, the language {a^nb^n | n 65 0} has the same Parikh image as (ab)^ 17. Roughly speaking, the theorem shows that if the order of the letters in a word is disregarded, retaining only the number of their occurrences, then context-free languages are indistinguishable from regular languages.
Due to the interesting theoretical property of the Parikh\u2019s theorem, the goal of this thesis is to study some aspects of descriptional complexity according to Parikh equivalence. In particular, we investigate the conversion of one-way nondeterministic \ufb01nite automata and context-free grammars into Parikh equivalent one-way and two-way deterministic \ufb01nite automata, from a descriptional complexity point of view.
We prove that for each one-way nondeterministic automaton with n states there exist Parikh equivalent one-way and two-way deterministic automata with e^O(sqrt(n lnn)) and p(n) states, respectively, where p(n) is a polynomial. Furthermore, these costs are tight. In contrast, if all the words accepted by the given one-way nondeterministic automaton contain at least two different letters, then a Parikh equivalent one-way deterministic automaton with a polynomial number of states can be found.
Concerning context-free grammars, we prove that for each grammar in Chomsky normal form with h variables there exist Parikh equivalent one-way and two-way deterministic automata with 2^O(h^2 ) and 2^O(h) states, respectively. Even these bounds are tight.
A further investigation is the study under Parikh equivalence of the state complexity of some language operations which preserve regularity. For union, concatenation, Kleene star, complement, intersection, shuffle, and reversal, we obtain a polynomial state complexity over any \ufb01xed alphabet, in contrast to the intrinsic exponential state complexity of some of these operations in the classical version. For projection we prove a superpolynomial state complexity, which is lower than the exponential one of the corresponding classical operation. We also prove that for each two one-way deterministic automata A and B it is possible to obtain a one-way deterministic automaton with a polynomial number of states whose accepted language has as Parikh image the intersection of the Parikh images of the languages accepted by A and B
Two-wayness: Automata and Transducers
This PhD is about two natural extensions of Finite Automata (FA): the 2-way fa (2FA) and the
2-way transducers (2T).
It is well known that 2FA s are computably equivalent to FAs, even in their nondeterministic
(2nfa) variant. However, in the field of descriptional complexity, some questions remain. Raised by Sakoda and Sipser in 1978, the question of the cost of the simulation of 2NFA by 2DFA (the deterministic variant of 2FA) is still open. In this manuscript, we give an answer in a restricted case in which the nondeterministic choices of the simulated 2NFA may occur at the boundaries of the input tape only (2ONFA). We show that every 2ONFA can be simulated by a 2DFA of subexponential (but superpolynomial) size. Under the assumptions L=NL, this cost is reduced to the polynomial level. Moreover, we prove that the complementation and the simulation by a halting 2ONFA is polynomial. We also consider the anologous simulations for alternating devices.
Providing a one-way write-only output tape to FAs leads to the notion of transducer. Contrary to the case of finite automata which are acceptor, 2-way transducers strictly extends the computational power of 1-way one, even in the case where both the input and output alphabets are unary. Though 1-way transducers enjoy nice properties and characterizations (algebraic, logical, etc. . . ), 2-way variants are less known, especially the nondeterministic case. In this area, this manuscript gives a new contribution: an algebraic characterization of the relations accepted by two-way transducers when both the input and output alphabets are unary. Actually, it can be reformulated as follows: each unary two-way transducer is equivalent to a sweeping (and even rotating) transducer. We also show that the assumptions made on the size of the alphabets are required, that is, sweeping transducers weakens the 2-way transducers whenever at least one of the alphabet is non-unary. On the path, we discuss on the computational power of some algebraic operations on word relations, introduced in the aim of describing the behavior of 2-way transducers or, more generally, of 2-way weighted automata. In particular, the mirror operation, consisting in reversing the input word in order to describe a right to left scan, draws our attention.
Finally, we study another kind of operations, more adapted for binary word relations: the
composition. We consider the transitive closure of relations. When the relation belongs to some very restricted sub-family of rational relations, we are able to compute its transitive closure and we set its complexity. This quickly becomes uncomputable when higher classes are considered