5,016 research outputs found
State Elimination Ordering Strategies: Some Experimental Results
Recently, the problem of obtaining a short regular expression equivalent to a
given finite automaton has been intensively investigated. Algorithms for
converting finite automata to regular expressions have an exponential blow-up
in the worst-case. To overcome this, simple heuristic methods have been
proposed.
In this paper we analyse some of the heuristics presented in the literature
and propose new ones. We also present some experimental comparative results
based on uniform random generated deterministic finite automata.Comment: In Proceedings DCFS 2010, arXiv:1008.127
Regular Combinators for String Transformations
We focus on (partial) functions that map input strings to a monoid such as
the set of integers with addition and the set of output strings with
concatenation. The notion of regularity for such functions has been defined
using two-way finite-state transducers, (one-way) cost register automata, and
MSO-definable graph transformations. In this paper, we give an algebraic and
machine-independent characterization of this class analogous to the definition
of regular languages by regular expressions. When the monoid is commutative, we
prove that every regular function can be constructed from constant functions
using the combinators of choice, split sum, and iterated sum, that are analogs
of union, concatenation, and Kleene-*, respectively, but enforce unique (or
unambiguous) parsing. Our main result is for the general case of
non-commutative monoids, which is of particular interest for capturing regular
string-to-string transformations for document processing. We prove that the
following additional combinators suffice for constructing all regular
functions: (1) the left-additive versions of split sum and iterated sum, which
allow transformations such as string reversal; (2) sum of functions, which
allows transformations such as copying of strings; and (3) function
composition, or alternatively, a new concept of chained sum, which allows
output values from adjacent blocks to mix.Comment: This is the full version, with omitted proofs and constructions, of
the conference paper currently in submissio
Small NFAs from Regular Expressions: Some Experimental Results
Regular expressions (res), because of their succinctness and clear syntax,
are the common choice to represent regular languages. However, efficient
pattern matching or word recognition depend on the size of the equivalent
nondeterministic finite automata (NFA). We present the implementation of
several algorithms for constructing small epsilon-free NFAss from res within
the FAdo system, and a comparison of regular expression measures and NFA sizes
based on experimental results obtained from uniform random generated res. For
this analysis, nonredundant res and reduced res in star normal form were
considered.Comment: Proceedings of 6th Conference on Computability in Europe (CIE 2010),
pages 194-203, Ponta Delgada, Azores, Portugal, June/July 201
Parametrized Stochastic Grammars for RNA Secondary Structure Prediction
We propose a two-level stochastic context-free grammar (SCFG) architecture
for parametrized stochastic modeling of a family of RNA sequences, including
their secondary structure. A stochastic model of this type can be used for
maximum a posteriori estimation of the secondary structure of any new sequence
in the family. The proposed SCFG architecture models RNA subsequences
comprising paired bases as stochastically weighted Dyck-language words, i.e.,
as weighted balanced-parenthesis expressions. The length of each run of
unpaired bases, forming a loop or a bulge, is taken to have a phase-type
distribution: that of the hitting time in a finite-state Markov chain. Without
loss of generality, each such Markov chain can be taken to have a bounded
complexity. The scheme yields an overall family SCFG with a manageable number
of parameters.Comment: 5 pages, submitted to the 2007 Information Theory and Applications
Workshop (ITA 2007
From Finite Automata to Regular Expressions and Back--A Summary on Descriptional Complexity
The equivalence of finite automata and regular expressions dates back to the
seminal paper of Kleene on events in nerve nets and finite automata from 1956.
In the present paper we tour a fragment of the literature and summarize results
on upper and lower bounds on the conversion of finite automata to regular
expressions and vice versa. We also briefly recall the known bounds for the
removal of spontaneous transitions (epsilon-transitions) on non-epsilon-free
nondeterministic devices. Moreover, we report on recent results on the average
case descriptional complexity bounds for the conversion of regular expressions
to finite automata and brand new developments on the state elimination
algorithm that converts finite automata to regular expressions.Comment: In Proceedings AFL 2014, arXiv:1405.527
- …