296 research outputs found
On Commutation and Conjugacy of Rational Languages and the Fixed Point Method
The research on language equations has been active during last decades. Compared to the equations on words the equations on languages are much more difficult to solve. Even very simple equations that are easy to solve for words can be very hard for languages. In this thesis we study two of such equations, namely commutation and conjugacy equations. We study these equations on some limited special cases and compare some of these results to the solutions of corresponding equations on words. For both equations we study the maximal solutions, the centralizer and the conjugator. We present a fixed point method that we can use to search these maximal solutions and analyze the reasons why this method is not successful for all languages. We give also several examples to illustrate the behaviour of this method.Siirretty Doriast
Optimal Qubit Mapping with Simultaneous Gate Absorption
Before quantum error correction (QEC) is achieved, quantum computers focus on
noisy intermediate-scale quantum (NISQ) applications. Compared to the
well-known quantum algorithms requiring QEC, like Shor's or Grover's algorithm,
NISQ applications have different structures and properties to exploit in
compilation. A key step in compilation is mapping the qubits in the program to
physical qubits on a given quantum computer, which has been shown to be an
NP-hard problem. In this paper, we present OLSQ-GA, an optimal qubit mapper
with a key feature of simultaneous SWAP gate absorption during qubit mapping,
which we show to be a very effective optimization technique for NISQ
applications. For the class of quantum approximate optimization algorithm
(QAOA), an important NISQ application, OLSQ-GA reduces depth by up to 50.0% and
SWAP count by 100% compared to other state-of-the-art methods, which translates
to 55.9% fidelity improvement. The solution optimality of OLSQ-GA is achieved
by the exact SMT formulation. For better scalability, we augment our approach
with additional constraints in the form of initial mapping or alternating
matching, which speeds up OLSQ-GA by up to 272X with no or little loss of
optimality.Comment: 8 pages, 8 figures, to appear in ICCAD'2
A Geometric Approach to the Problem of Unique Decomposition of Processes
This paper proposes a geometric solution to the problem of prime
decomposability of concurrent processes first explored by R. Milner and F.
Moller in [MM93]. Concurrent programs are given a geometric semantics using
cubical areas, for which a unique factorization theorem is proved. An effective
factorization method which is correct and complete with respect to the
geometric semantics is derived from the factorization theorem. This algorithm
is implemented in the static analyzer ALCOOL.Comment: 15 page
On the k-Abelian Equivalence Relation of Finite Words
This thesis is devoted to the so-called k-abelian equivalence relation of sequences of symbols, that is, words. This equivalence relation is a generalization of the abelian equivalence of words. Two words are abelian equivalent if one is a permutation of the other. For any positive integer k, two words are called k-abelian equivalent if each word of length at most k occurs equally many times as a factor in the two words. The k-abelian equivalence defines an equivalence relation, even a congruence, of finite words. A hierarchy of equivalence classes in between the equality relation and the abelian equivalence of words is thus obtained.
Most of the literature on the k-abelian equivalence deals with infinite words. In this thesis we consider several aspects of the equivalence relations, the main objective being to build a fairly comprehensive picture on the structure of the k-abelian equivalence classes themselves. The main part of the thesis deals with the structural aspects of k-abelian equivalence classes. We also consider aspects of k-abelian equivalence in infinite words.
We survey known characterizations of the k-abelian equivalence of finite words from the literature and also introduce novel characterizations. For the analysis of structural properties of the equivalence relation, the main tool is the characterization by the rewriting rule called the k-switching. Using this rule it is straightforward to show that the language comprised of the lexicographically least elements of the k-abelian equivalence classes is regular. Further word-combinatorial analysis of the lexicographically least elements leads us to describe the deterministic finite automata recognizing this language. Using tools from formal language theory combined with our analysis, we give an optimal expression for the asymptotic growth rate of the number of k-abelian equivalence classes of length n over an m-letter alphabet. Explicit formulae are computed for small values of k and m, and these sequences appear in Sloane’s Online Encyclopedia of Integer Sequences.
Due to the fact that the k-abelian equivalence relation is a congruence of the free monoid, we study equations over the k-abelian equivalence classes. The main result in this setting is that any system of equations of k-abelian equivalence classes is equivalent to one of its finite subsystems, i.e., the monoid defined by the k-abelian equivalence relation possesses the compactness property.
Concerning infinite words, we mainly consider the (k-)abelian complexity function. We complete a classification of the asymptotic abelian complexities of pure morphic binary words. In other words, given a morphism which has an infinite binary fixed point, the limit superior asymptotic abelian complexity of the fixed point can be computed (in principle). We also give a new proof of the fact that the k-abelian complexity of a Sturmian word is n + 1 for length n 2k. In fact, we consider several aspects of the k-abelian equivalence relation in Sturmian words using a dynamical interpretation of these words. We reprove the fact that any Sturmian word contains arbitrarily large k-abelian repetitions. The methods used allow to analyze the situation in more detail, and this leads us to define the so-called k-abelian critical exponent which measures the ratio of the exponent and the length of the root of a k-abelian repetition. This notion is connected to a deep number theoretic object called the Lagrange spectrum
Termination of Narrowing: Automated Proofs and Modularity Properties
En 1936 Alan Turing demostro que el halting problem, esto es, el problema de decidir
si un programa termina o no, es un problema indecidible para la inmensa mayoria de
los lenguajes de programacion. A pesar de ello, la terminacion es un problema tan
relevante que en las ultimas decadas un gran numero de tecnicas han sido desarrolladas
para demostrar la terminacion de forma automatica de la maxima cantidad posible de
programas. Los sistemas de reescritura de terminos proporcionan un marco teorico
abstracto perfecto para el estudio de la terminacion de programas. En este marco, la
evaluaci on de un t ermino consiste en la aplicacion no determinista de un conjunto de
reglas de reescritura.
El estrechamiento (narrowing) de terminos es una generalizacion de la reescritura
que proporciona un mecanismo de razonamiento automatico. Por ejemplo, dado un
conjunto de reglas que denan la suma y la multiplicacion, la reescritura permite calcular
expresiones aritmeticas, mientras que el estrechamiento permite resolver ecuaciones
con variables. Esta tesis constituye el primer estudio en profundidad de las
propiedades de terminacion del estrechamiento. Las contribuciones son las siguientes.
En primer lugar, se identican clases de sistemas en las que el estrechamiento tiene
un comportamiento bueno, en el sentido de que siempre termina. Muchos metodos
de razonamiento automatico, como el analisis de la semantica de lenguajes de programaci
on mediante operadores de punto jo, se benefician de esta caracterizacion.
En segundo lugar, se introduce un metodo automatico, basado en el marco teorico
de pares de dependencia, para demostrar la terminacion del estrechamiento en un
sistema particular. Nuestro metodo es, por primera vez, aplicable a cualquier clase
de sistemas.
En tercer lugar, se propone un nuevo metodo para estudiar la terminacion del
estrechamiento desde un termino particular, permitiendo el analisis de la terminacion
de lenguajes de programacion. El nuevo metodo generaliza losIborra López, J. (2010). Termination of Narrowing: Automated Proofs and Modularity Properties [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/19251Palanci
Probabilistic call by push value
We introduce a probabilistic extension of Levy's Call-By-Push-Value. This
extension consists simply in adding a " flipping coin " boolean closed atomic
expression. This language can be understood as a major generalization of
Scott's PCF encompassing both call-by-name and call-by-value and featuring
recursive (possibly lazy) data types. We interpret the language in the
previously introduced denotational model of probabilistic coherence spaces, a
categorical model of full classical Linear Logic, interpreting data types as
coalgebras for the resource comonad. We prove adequacy and full abstraction,
generalizing earlier results to a much more realistic and powerful programming
language
Routines and Applications of Symbolic Algebra Software
Computing has become an essential resource in modern research and has found application
across a wide range of scientific disciplines. Developments in symbolic algebra tools have been
particularly valuable in physics where calculations in fields such as general relativity, quantum
field theory and physics beyond the standard model are becoming increasing complex and
unpractical to work with by hand. The computer algebra system Cadabra is a tensor-first
approach to symbolic algebra based on the programming language Python which has been used
extensively in research in these fields while also having a shallow learning curve making it an
excellent way to introduce students to methods in computer algebra.
The work in this thesis has been concentrated on developing Cadabra, which has involved
looking at two different elements which make up a computer algebra program. Firstly, the
implementation of algebraic routines is discussed. This has primarily been focused on the
introduction of an algorithm for detecting the equivalence of tensorial expressions related by
index permutation symmetries. The method employed differs considerably from traditional
canonicalisation routines which are commonly used for this purpose by using Young projection
operators to make such symmetries manifest.
The other element of writing a computer algebra program which is covered is the infrastruc-
ture and environment. The importance of this aspect of software design is often overlooked by
funding committees and academic software users resulting in an anti-pattern of code not being
shared and contributed to in the way in which research itself is published and promulgated.
The focus in this area has been on implementing a packaging system for Cadabra which allows
the writing of generic libraries which can be shared by the community, and interfacing with
other scientific computing packages to increase the capabilities of Cadabra
Full abstraction for probabilistic PCF
We present a probabilistic version of PCF, a well-known simply typed
universal functional language. The type hierarchy is based on a single ground
type of natural numbers. Even if the language is globally call-by-name, we
allow a call-by-value evaluation for ground type arguments in order to provide
the language with a suitable algorithmic expressiveness. We describe a
denotational semantics based on probabilistic coherence spaces, a model of
classical Linear Logic developed in previous works. We prove an adequacy and an
equational full abstraction theorem showing that equality in the model
coincides with a natural notion of observational equivalence
- …