626 research outputs found
Cyclic Datatypes modulo Bisimulation based on Second-Order Algebraic Theories
Cyclic data structures, such as cyclic lists, in functional programming are
tricky to handle because of their cyclicity. This paper presents an
investigation of categorical, algebraic, and computational foundations of
cyclic datatypes. Our framework of cyclic datatypes is based on second-order
algebraic theories of Fiore et al., which give a uniform setting for syntax,
types, and computation rules for describing and reasoning about cyclic
datatypes. We extract the "fold" computation rules from the categorical
semantics based on iteration categories of Bloom and Esik. Thereby, the rules
are correct by construction. We prove strong normalisation using the General
Schema criterion for second-order computation rules. Rather than the fixed
point law, we particularly choose Bekic law for computation, which is a key to
obtaining strong normalisation. We also prove the property of "Church-Rosser
modulo bisimulation" for the computation rules. Combining these results, we
have a remarkable decidability result of the equational theory of cyclic data
and fold.Comment: 38 page
Inferring Argument Size Relationships with CLP(R)
Argument size relationships are useful in termination analysis which, in turn, is important in program synthesis and goal-replacement transformations. We show how a precise analysis for inter-argument size relationships, formulated in terms of abstract interpretation, can be implemented straightforwardly in a language with constraint support like CLP(R) or SICStus version 3. The analysis is based on polyhedral approximations and uses a simple relaxation technique to calculate least upper bounds and a delay method to improve the precision of widening. To the best of our knowledge, and despite its simplicity, the analysis derives relationships to an accuracy that is either comparable or better than any existing technique
Effective lambda-models vs recursively enumerable lambda-theories
A longstanding open problem is whether there exists a non syntactical model
of the untyped lambda-calculus whose theory is exactly the least lambda-theory
(l-beta). In this paper we investigate the more general question of whether the
equational/order theory of a model of the (untyped) lambda-calculus can be
recursively enumerable (r.e. for brevity). We introduce a notion of effective
model of lambda-calculus calculus, which covers in particular all the models
individually introduced in the literature. We prove that the order theory of an
effective model is never r.e.; from this it follows that its equational theory
cannot be l-beta or l-beta-eta. We then show that no effective model living in
the stable or strongly stable semantics has an r.e. equational theory.
Concerning Scott's semantics, we investigate the class of graph models and
prove that no order theory of a graph model can be r.e., and that there exists
an effective graph model whose equational/order theory is minimum among all
theories of graph models. Finally, we show that the class of graph models
enjoys a kind of downwards Lowenheim-Skolem theorem.Comment: 34
A Multi-Threading Algorithm for Constrained Path Optimization Problem on Road Networks
The constrained path optimization (CPO) problem takes the following input:
(a) a road network represented as a directed graph, where each edge is
associated with a "cost" and a "score" value; (b) a source-destination pair
and; (c) a budget value, which denotes the maximum permissible cost of the
solution. Given the input, the goal is to determine a path from source to
destination, which maximizes the "score" while constraining the total "cost" of
the path to be within the given budget value. CPO problem has applications in
urban navigation. However, the CPO problem is computationally challenging as it
can be reduced to an instance of the arc orienteering problem, which is known
to be NP-hard. The current state-of-the-art algorithms for this problem are
essentially serial in nature and cannot take full advantage (i.e., achieve good
load balance) of the increasingly available multi-core systems to solve a CPO
query. Our proposed parallel algorithm (with its intelligent task-assignment
scheme) achieves both superior solution quality and very low execution times
(via good load balancing). Moreover, our approach is also able to demonstrate
an almost linear speed-up with an increase in the number of cores.Comment: 10 pages, 14 figures, accepted as a short paper in the 23rd
International Conference on Web Information Systems Engineerin
The Rice-Shapiro theorem in Computable Topology
We provide requirements on effectively enumerable topological spaces which
guarantee that the Rice-Shapiro theorem holds for the computable elements of
these spaces. We show that the relaxation of these requirements leads to the
classes of effectively enumerable topological spaces where the Rice-Shapiro
theorem does not hold. We propose two constructions that generate effectively
enumerable topological spaces with particular properties from wn--families and
computable trees without computable infinite paths. Using them we propose
examples that give a flavor of this class
Categorial Compositionality III: F-(co)algebras and the Systematicity of Recursive Capacities in Human Cognition
Human cognitive capacity includes recursively definable concepts, which are prevalent in domains involving lists, numbers, and languages. Cognitive science currently lacks a satisfactory explanation for the systematic nature of such capacities (i.e., why the capacity for some recursive cognitive abilities–e.g., finding the smallest number in a list–implies the capacity for certain others–finding the largest number, given knowledge of number order). The category-theoretic constructs of initial F-algebra, catamorphism, and their duals, final coalgebra and anamorphism provide a formal, systematic treatment of recursion in computer science. Here, we use this formalism to explain the systematicity of recursive cognitive capacities without ad hoc assumptions (i.e., to the same explanatory standard used in our account of systematicity for non-recursive capacities). The presence of an initial algebra/final coalgebra explains systematicity because all recursive cognitive capacities, in the domain of interest, factor through (are composed of) the same component process. Moreover, this factorization is unique, hence no further (ad hoc) assumptions are required to establish the intrinsic connection between members of a group of systematically-related capacities. This formulation also provides a new perspective on the relationship between recursive cognitive capacities. In particular, the link between number and language does not depend on recursion, as such, but on the underlying functor on which the group of recursive capacities is based. Thus, many species (and infants) can employ recursive processes without having a full-blown capacity for number and language
LCF Examples in HOL
The LCF system provides a logic of fixed point theory and is useful to reason about non-termination, arbitrary recursive definitions and infinite types as lazy lists. It is unsuitable for reasoning about finite types and strict functions. The HOL system provides set theory and supports reasoning about finite types and total functions well. In this paper a number of examples are used to demonstrate that an extension of HOL with domain theory combines the benefits of both systems. The examples illustrate reasoning about infinite values and non-terminating functions and show how mixing domain and set theoretic reasoning eases reasoning about finite LCF types and strict functions. An example presents a proof of the correctness and termination of a recursive unification algorithm using well-founded induction
Compound poisson approximation via information functionals
An information-theoretic development is given for the problem of compound Poisson approximation, which parallels earlier treatments for Gaussian and Poisson approximation. Nonasymptotic bounds are derived for the distance between the distribution of a sum of independent integer-valued random variables and an appropriately chosen compound Poisson law. In the case where all summands have the same conditional distribution given that they are non-zero, a bound on the relative entropy distance between their sum and the compound Poisson distribution is derived, based on the data-processing property of relative entropy and earlier Poisson approximation results. When the summands have arbitrary distributions, corresponding bounds are derived in terms of the total variation distance. The main technical ingredient is the introduction of two "information functionals,'' and the analysis of their properties. These information functionals play a role analogous to that of the classical Fisher information in normal approximation. Detailed comparisons are made between the resulting inequalities and related bounds
- …