26,046 research outputs found

    Towards Intelligent Databases

    Get PDF
    This article is a presentation of the objectives and techniques of deductive databases. The deductive approach to databases aims at extending with intensional definitions other database paradigms that describe applications extensionaUy. We first show how constructive specifications can be expressed with deduction rules, and how normative conditions can be defined using integrity constraints. We outline the principles of bottom-up and top-down query answering procedures and present the techniques used for integrity checking. We then argue that it is often desirable to manage with a database system not only database applications, but also specifications of system components. We present such meta-level specifications and discuss their advantages over conventional approaches

    The ERA of FOLE: Superstructure

    Full text link
    This paper discusses the representation of ontologies in the first-order logical environment FOLE (Kent 2013). An ontology defines the primitives with which to model the knowledge resources for a community of discourse (Gruber 2009). These primitives, consisting of classes, relationships and properties, are represented by the ERA (entity-relationship-attribute) data model (Chen 1976). An ontology uses formal axioms to constrain the interpretation of these primitives. In short, an ontology specifies a logical theory. This paper is the second in a series of three papers that provide a rigorous mathematical representation for the ERA data model in particular, and ontologies in general, within the first-order logical environment FOLE. The first two papers show how FOLE represents the formalism and semantics of (many-sorted) first-order logic in a classification form corresponding to ideas discussed in the Information Flow Framework (IFF). In particular, the first paper (Kent 2015) provided a "foundation" that connected elements of the ERA data model with components of the first-order logical environment FOLE, and this second paper provides a "superstructure" that extends FOLE to the formalisms of first-order logic. The third paper will define an "interpretation" of FOLE in terms of the transformational passage, first described in (Kent 2013), from the classification form of first-order logic to an equivalent interpretation form, thereby defining the formalism and semantics of first-order logical/relational database systems (Kent 2011). The FOLE representation follows a conceptual structures approach, that is completely compatible with Formal Concept Analysis (Ganter and Wille 1999) and Information Flow (Barwise and Seligman 1997)

    Arnošt Kolman’s Critique of Mathematical Fetishism

    Get PDF
    Arnošt Kolman (1892–1979) was a Czech mathematician, philosopher and Communist official. In this paper, we would like to look at Kolman’s arguments against logical positivism which revolve around the notion of the fetishization of mathematics. Kolman derives his notion of fetishism from Marx’s conception of commodity fetishism. Kolman is aiming to show the fact that an entity (system, structure, logical construction) acquires besides its real existence another formal existence. Fetishism means the fantastic detachment of the physical characteristics of real things or phenomena from these things. We identify Kolman’s two main arguments against logical positivism. In the first argument, Kolman applied Lenin’s arguments against Mach’s empiricism-criticism onto Russell’s neutral monism, i.e. mathematical fetishism is internally related to political conservativism. Kolman’s second main argument is that logical and mathematical fetishes are epistemologically deprived of any historical and dynamic dimension. In the final parts of our paper we place Kolman’s thinking into the context of his time, and furthermore we identify some tenets of mathematical fetishism appearing in Alain Badiou’s mathematical ontology today

    An eclectic quadrant of rule based system verification: work grounded in verification of fuzzy rule bases.

    Get PDF
    In this paper, we used a research approach based on grounded theory in order to classify methods proposed in literature that try to extend the verification of classical rule bases to the case of fuzzy knowledge modeling. Within this area of verification we identify two dual lines of thought respectively leading to what is termed respectively static and dynamic anomaly detection methods. The major outcome of the confrontation of both approaches is that their results, most often stated in terms of necessary and/or sufficient conditions are difficult to reconcile. This paper addresses precisely this issue by the construction of a theoretical framework, which enables to effectively evaluate the results of both static and dynamic verification theories. Things essentially go wrong when in the quest for a good affinity, matching or similarity measure, one neglects to take into account the effect of the implication operator, an issue that rises above and beyond the fuzzy setting that initiated the research. The findings can easily be generalized to verification issues in any knowledge coding setting.Systems;
    • …
    corecore