56 research outputs found

    Formal Concept Analysis Methods for Description Logics

    Get PDF
    This work presents mainly two contributions to Description Logics (DLs) research by means of Formal Concept Analysis (FCA) methods: supporting bottom-up construction of DL knowledge bases, and completing DL knowledge bases. Its contribution to FCA research is on the computational complexity of computing generators of closed sets

    On the Computation of Common Subsumers in Description Logics

    Get PDF
    Description logics (DL) knowledge bases are often build by users with expertise in the application domain, but little expertise in logic. To support this kind of users when building their knowledge bases a number of extension methods have been proposed to provide the user with concept descriptions as a starting point for new concept definitions. The inference service central to several of these approaches is the computation of (least) common subsumers of concept descriptions. In case disjunction of concepts can be expressed in the DL under consideration, the least common subsumer (lcs) is just the disjunction of the input concepts. Such a trivial lcs is of little use as a starting point for a new concept definition to be edited by the user. To address this problem we propose two approaches to obtain "meaningful" common subsumers in the presence of disjunction tailored to two different methods to extend DL knowledge bases. More precisely, we devise computation methods for the approximation-based approach and the customization of DL knowledge bases, extend these methods to DLs with number restrictions and discuss their efficient implementation

    Computing Least Common Subsumers in ALEN

    Get PDF
    Computing the least common subsumer (lcs) in description logics is an inference task first introduced for sublanguages of CLASSIC. Roughly speaking, the lcs of a set of concept descriptions is the most specific concept description that subsumes all of the input descriptions. As such, the lcs allows to extract the commonalities from given concept descriptions, a task essential for several applications like, e.g., inductive learning, information retrieval, or the bottom-up construction of KR-knowledge bases. Previous work on the lcs has concentrated on description logics that either allow for number restrictions or for existential restrictions. Many applications, however, require to combine these constructors. In this work, we present an lcs algorithm for the description logic ALEN, which allows for both constructors (as well as concept conjunction, primitive negation, and value restrictions). The proof of correctness of our lcs algorithm is based on an appropriate structural characterization of subsumption in ALEN also introduced in this paper.This research was carried out while the second author was still at the LuFG Theoretical Computer Science, RWTH Aachen

    Microtheories for SDI - Accounting for diversity of local conceptualisations at a global level

    Get PDF
    Dissertation submitted in partial fulfillment of the requirements for the Degree of Master of Science in Geospatial Technologies.The categorization and conceptualization of geographic features is fundamental to cartography, geographic information retrieval, routing applications, spatial decision support and data sharing in general. However, there is no standard conceptualization of the world. Humans conceptualize features based on numerous factors including cultural background, knowledge, motivation and particularly space and time. Thus, geographic features are prone to multiple, context-dependent conceptualizations reflecting local conditions. This creates semantic heterogeneity and undermines interoperability. Standardization of a shared definition is often employed to overcome semantic heterogeneity. However, this approach loses important local diversity in feature conceptualizations and may result in feature definitions which are too broad or too specific. This work proposes the use of microtheories in Spatial Data Infrastructures, such as INSPIRE, to account for diversity of local conceptualizations while maintaining interoperability at a global level. It introduces a novel method of structuring microtheories based on space and time, represented by administrative boundaries, to reflect variations in feature conceptualization. A bottom-up approach, based on non-standard inference, is used to create an appropriate global-level feature definition from the local definitions. Conceptualizations of rivers, forests and estuaries throughout Europe are used to demonstrate how the approach can improve the INSPIRE data model and ease its adoption by European member states

    Reasoning-Supported Quality Assurance for Knowledge Bases

    Get PDF
    The increasing application of ontology reuse and automated knowledge acquisition tools in ontology engineering brings about a shift of development efforts from knowledge modeling towards quality assurance. Despite the high practical importance, there has been a substantial lack of support for ensuring semantic accuracy and conciseness. In this thesis, we make a significant step forward in ontology engineering by developing a support for two such essential quality assurance activities

    Learning Description Logic Knowledge Bases from Data Using Methods from Formal Concept Analysis

    Get PDF
    Description Logics (DLs) are a class of knowledge representation formalisms that can represent terminological and assertional knowledge using a well-defined semantics. Often, knowledge engineers are experts in their own fields, but not in logics, and require assistance in the process of ontology design. This thesis presents three methods that can extract terminological knowledge from existing data and thereby assist in the design process. They are based on similar formalisms from Formal Concept Analysis (FCA), in particular the Next-Closure Algorithm and Attribute-Exploration. The first of the three methods computes terminological knowledge from the data, without any expert interaction. The two other methods use expert interaction where a human expert can confirm each terminological axiom or refute it by providing a counterexample. These two methods differ only in the way counterexamples are provided

    A knowledge acquisition assistant for the expert system shell Nexpert-Object

    Get PDF
    This study addresses the problems of knowledge acquisition in expert system development examines programs whose goal is to solve part of these problems. Among them are knowledge acquisition tools, which provide the knowledge engineer with a set of Artificial Intelligence primitives, knowledge acquisition aids, which offer to the knowledge engineer a guidance in knowledge elicitation, and finally, automated systems, which try to replace the human interviewer with a machine interface. We propose an alternative technique to these approaches: an interactive syntactic analyzer of an emerging knowledge base written with the expert system shell called Nexpert Object. This program intends to help the knowledge engineer during the editing of a knowledge base, both from a knowledge engineering and a knowledge representation point of view. The implementation is a Desk Accessory written in C, running on Macintosh concurrently with Nexpert Object

    Chord sequence patterns in OWL

    Get PDF
    This thesis addresses the representation of, and reasoning on, musical knowledge in the Semantic Web. The Semantic Web is an evolving extension of the World Wide Web that aims at describing information that is distributed on the web in a machine-processable form. Existing approaches to modelling musical knowledge in the context of the Semantic Web have focused on metadata. The description of musical content and reasoning as well as integration of content descriptions and metadata are yet open challenges. This thesis discusses the possibilities of representing musical knowledge in the Web Ontology Language (OWL) focusing on chord sequence representation and presents and evaluates a newly developed solution. The solution consists of two main components. Ontological modelling patterns for musical entities such as notes and chords are introduced in the (MEO) ontology. A sequence pattern language and ontology (SEQ) has been developed that can express patterns in a form resembling regular expressions. As MEO and SEQ patterns both rewrite to OWL they can be combined freely. Reasoning tasks such as instance classification, retrieval and pattern subsumption are then executable by standard Semantic Web reasoners. The expressiveness of SEQ has been studied, in particular in relation to grammars. The complexity of reasoning on SEQ patterns has been studied theoretically and empirically, and optimisation methods have been developed. There is still great potential for improvement if specific reasoning algorithms were developed to exploit the sequential structure, but the development of such algorithms is outside the scope of this thesis. MEO and SEQ have also been evaluated in several musicological scenarios. It is shown how patterns that are characteristic of musical styles can be expressed and chord sequence data can be classified, demonstrating the use of the language in web retrieval and as integration layer for different chord patterns and corpora. Furthermore, possibilities of using SEQ patterns for harmonic analysis are explored using grammars for harmony; both a hybrid system and a translation of limited context-free grammars into SEQ patterns have been developed. Finally, a distributed scenario is evaluated where SEQ and MEO are used in connection with DBpedia, following the Linked Data approach. The results show that applications are already possible and will benefit in the future from improved quality and compatibility of data sources as the Semantic Web evolves.EThOS - Electronic Theses Online ServiceGBUnited Kingdo
    • …
    corecore