8 research outputs found

    Canonical Horn Representations and Query Learning

    Get PDF
    We describe an alternative construction of an existing canonical representation for definite Horn theories, the Guigues-Duquenne basis (or GD basis), which minimizes a natural notion of implicational size. We extend the canonical representation to general Horn, by providing a reduction from definite to general Horn CNF. We show how this representation relates to two topics in query learning theory: first, we show that a well-known algorithm by Angluin, Frazier and Pitt that learns Horn CNF always outputs the GD basis independently of the counterexamples it receives; second, we build strong polynomial certificates for Horn CNF directly from the GD basis

    Abstract Concept Lattices

    No full text
    International audienceWe present a view of abstraction based on a structure preserving reduction of the Galois connection between a language of terms and the powerset of a set of instances O. Such a relation is materialized as an extension-intension lattice, namely a concept lattice when L is the powerset of a set P of attributes. We define and characterize an abstraction A as some part of either the language or the powerset of O, defined in such a way that the extension-intension latticial structure is preserved. Such a structure is denoted for short as an abstract lattice. We discuss the extensional abstract lattices obtained by so reducing the powerset of O, together together with the corresponding abstract implications, and discuss alpha lattices as particular abstract lattices. Finally we give formal framework allowing to define a generalized abstract lattice whose language is made of terms mixing abstract and non abstract conjunctions of properties

    On the estimation of cointegration models

    No full text
    Includes bibliographical referencesAvailable from British Library Document Supply Centre- DSC:DX221226 / BLDSC - British Library Document Supply CentreSIGLEGBUnited Kingdo
    corecore