212,944 research outputs found

    On the Triplet Frame for Concept Analysis

    Get PDF
    The paper has two objectives: to introduce the fundamentals of a triplet model of a concept, and to show that the main concept models may be structurally treated as its partial cases. The triplet model considers a concept as a mental representation and characterizes it from three interrelated perspectives. The first deals with objects (and their attributes of various orders) subsumed under a concept. The second focuses on representing structures that depict objects and their attributes in some intelligent system. The third concentrates on the ways of establishing correspondences between objects with their attributes and appropriate representing structures

    Representing Concepts by Weighted Formulas

    Get PDF
    A concept is traditionally defined via the necessary and sufficient conditions that clearly determine its extension. By contrast, cognitive views of concepts intend to account for empirical data that show that categorisation under a concept presents typicality effects and a certain degree of indeterminacy. We propose a formal language to compactly represent concepts by leveraging on weighted logical formulas. In this way, we can model the possible synergies among the qualities that are relevant for categorising an object under a concept. We show that our proposal can account for a number of views of concepts such as the prototype theory and the exemplar theory. Moreover, we show how the proposed model can overcome some limitations of cognitive views

    A Logic-based Approach for Recognizing Textual Entailment Supported by Ontological Background Knowledge

    Full text link
    We present the architecture and the evaluation of a new system for recognizing textual entailment (RTE). In RTE we want to identify automatically the type of a logical relation between two input texts. In particular, we are interested in proving the existence of an entailment between them. We conceive our system as a modular environment allowing for a high-coverage syntactic and semantic text analysis combined with logical inference. For the syntactic and semantic analysis we combine a deep semantic analysis with a shallow one supported by statistical models in order to increase the quality and the accuracy of results. For RTE we use logical inference of first-order employing model-theoretic techniques and automated reasoning tools. The inference is supported with problem-relevant background knowledge extracted automatically and on demand from external sources like, e.g., WordNet, YAGO, and OpenCyc, or other, more experimental sources with, e.g., manually defined presupposition resolutions, or with axiomatized general and common sense knowledge. The results show that fine-grained and consistent knowledge coming from diverse sources is a necessary condition determining the correctness and traceability of results.Comment: 25 pages, 10 figure

    Probabilistic Programming Concepts

    Full text link
    A multitude of different probabilistic programming languages exists today, all extending a traditional programming language with primitives to support modeling of complex, structured probability distributions. Each of these languages employs its own probabilistic primitives, and comes with a particular syntax, semantics and inference procedure. This makes it hard to understand the underlying programming concepts and appreciate the differences between the different languages. To obtain a better understanding of probabilistic programming, we identify a number of core programming concepts underlying the primitives used by various probabilistic languages, discuss the execution mechanisms that they require and use these to position state-of-the-art probabilistic languages and their implementation. While doing so, we focus on probabilistic extensions of logic programming languages such as Prolog, which have been developed since more than 20 years

    The Mode of Computing

    Full text link
    The Turing Machine is the paradigmatic case of computing machines, but there are others, such as Artificial Neural Networks, Table Computing, Relational-Indeterminate Computing and diverse forms of analogical computing, each of which based on a particular underlying intuition of the phenomenon of computing. This variety can be captured in terms of system levels, re-interpreting and generalizing Newell's hierarchy, which includes the knowledge level at the top and the symbol level immediately below it. In this re-interpretation the knowledge level consists of human knowledge and the symbol level is generalized into a new level that here is called The Mode of Computing. Natural computing performed by the brains of humans and non-human animals with a developed enough neural system should be understood in terms of a hierarchy of system levels too. By analogy from standard computing machinery there must be a system level above the neural circuitry levels and directly below the knowledge level that is named here The mode of Natural Computing. A central question for Cognition is the characterization of this mode. The Mode of Computing provides a novel perspective on the phenomena of computing, interpreting, the representational and non-representational views of cognition, and consciousness.Comment: 35 pages, 8 figure
    • …
    corecore