24 research outputs found

    Game semantics approach to higher-order complexity

    Get PDF
    Game semantics was initially defined and used to characterize pcf functionals. We use this approach to propose a definition of complexity for such higher-order functions, as well as a class of polynomial time computable higher-order functions

    Mechanised Uniform Interpolation for Modal Logics K, GL, and iSL

    Get PDF
    The uniform interpolation property in a given logic can be understood as the definability of propositional quantifiers. We mechanise the computation of these quantifiers and prove correctness in the Coq proof assistant for three modal logics, namely: (1) the modal logic K, for which a pen-and-paper proof exists; (2) Gödel-Löb logic GL, for which our formalisation clarifies an important point in an existing, but incomplete, sequent-style proof; and (3) intuitionistic strong Löb logic iSL, for which this is the first proof-theoretic construction of uniform interpolants. Our work also yields verified programs that allow one to compute the propositional quantifiers on any formula in this logic

    Mechanised Uniform Interpolation for Modal Logics K, GL, and iSL

    Get PDF
    The uniform interpolation property in a given logic can be understood as the definability of propositional quantifiers. We mechanise the computation of these quantifiers and prove correctness in the Coq proof assistant for three modal logics, namely: (1) the modal logic K, for which a pen-and-paper proof exists; (2) Gödel-Löb logic GL, for which our formalisation clarifies an important point in an existing, but incomplete, sequent-style proof; and (3) intuitionistic strong Löb logic iSL, for which this is the first proof-theoretic construction of uniform interpolants. Our work also yields verified programs that allow one to compute the propositional quantifiers on any formula in this logic

    Probing for Surface Mesh Generation through Delaunay Refinement

    Get PDF
    Surface mesh generation through Delaunay refinement is considered as a relevant alternative to the common marching cubes algorithm. One distinctive feature lies into the fact that it interleaves shape probing (through intersection with Voronoi edges) with refinement. The current implementations require seeding the refinement procedure for each connected component so as to guarantee that all components are properly discovered by the mesh refinement procedure. This task is often left to the user. Although this is an easy task for input polyhedral surfaces, it is not when the input surface is defined as an isovalue of an implicit function. In this report we propose an automatic seeding procedure which interleaves refinement, seeding and initialization steps with the qualities of the refinement and the guarantees of careful seeding

    Higher-order complexity in analysis

    Get PDF
    International audienceWe present ongoing work on the development of complexity theory in analysis. Kawamura and Cook recently showed how to carry out complexity theory on the space C[0,1] of continuous real functions on the unit interval. It is done, as in computable analysis, by representing objects by first-order functions (from finite words to finite words, say) and by measuring the complexity of a second-order functional in terms of second-order polynomials. We prove that this framework cannot be directly applied to spaces that are not σ\sigma-compact. However, representing objects by higher-order functions (over finite words, say) makes it possible to carry out complexity theory on such spaces: for this purpose we develop the complexity of higher-order functionals. At orders above 3, our class of polynomial-time computable functionals strictly contains the class BFF of Buss, Cook and Urquhart

    ROTOR: A Tool for Renaming Values in OCaml’s Module System

    Get PDF
    The functional programming paradigm presents its own unique challenges to refactoring. For the OCaml language in particular, the expressiveness of its module system makes this a highly non-trivial task and there is currently no automated support for large-scale refactoring in the OCaml language. We present ROTOR, a tool for automatically renaming toplevel value definitions in OCaml’s module system. To compute the effect of renaming, ROTOR relies on a novel concept which we call a value extension. This is a collection of related declarationsin a program that must all be renamed at once. In practice, this leads to a notion of dependency: renaming a function foo in module A (mutually) depends on renaming function foo in module B etc. We describe important aspects of ROTOR’s design, implementation, and evaluation on two large codebases: Jane Street’s core library and its dependencies, and the OCaml compiler itself. In these real-world settings we find that some cases involve a surprisingly complex network of dependencies, and that the use of the PPX preprocessor system presents significant challenges

    On the query complexity of real functionals

    Get PDF
    Recently Kawamura and Cook developed a framework to define the computational complexity of operators arising in analysis. Our goal is to understand the effects of complexity restrictions on the analytical properties of the operator. We focus on the case of norms over C[0,1] and introduce the notion of dependence of a norm on a point and relate it to the query complexity of the norm. We show that the dependence of almost every point is of the order of the query complexity of the norm. A norm with small complexity depends on a few points but, as compensation, highly depends on them. We characterize the functionals that are computable using one oracle call only and discuss the uniformity of that characterization

    Characterizing polynomial time complexity of stream programs using interpretations

    Get PDF
    This paper provides a criterion based on interpretation methods on term rewrite systems in order to characterize the polynomial time complexity of second order functionals. For that purpose it introduces a first order functional stream language that allows the programmer to implement second order functionals. This characterization is extended through the use of exp-poly interpretations as an attempt to capture the class of Basic Feasible Functionals (bff). Moreover, these results are adapted to provide a new characterization of polynomial time complexity in computable analysis. These characterizations give a new insight on the relations between the complexity of functional stream programs and the classes of functions computed by Oracle Turing Machine, where oracles are treated as inputs

    Analytical properties of resource-bounded real functionals

    Get PDF
    Computable analysis is an extension of classical discrete computability by enhancing the normal Turing machine model. It investigates mathematical analysis from the computability perspective. Though it is well developed on the computability level, it is still under developed on the complexity perspective, that is, when bounding the available computational resources. Recently Kawamura and Cook developed a framework to define the computational complexity of operators arising in analysis. Our goal is to understand the effects of complexity restrictions on the analytical properties of the operator. We focus on the case of norms over C[0,1] and introduce the notion of dependence of a norm on a point and relate it to the query complexity of the norm. We show that the dependence of almost every point is of the order of the query complexity of the norm. A norm with small complexity depends on a few points but, as compensation, highly depends on them. We briefly show how to obtain similar results for non-deterministic time complexity. We characterize the functionals that are computable using one oracle call only and discuss the uniformity of that characterization. This paper is a significant revision and expansion of an earlier conference version
    corecore