12 research outputs found

    Minimizing the Estimated Solution Cost with A* Search to Support Minimal Mapping Repair

    Get PDF
    Incoherent alignment has been the main focus in the matching process since 2010.  Incoherent means that there is semantic or logic conflict in the alignment. This condition encouraged researches in ontology matching field to improve the alignment by repairing the incoherent alignment. Repair mapping will restore the incoherent to coherent mapping, by deleting unwanted mappings from the alignment. In order to minimize the impacts in the input alignment, repair process should be done as as minimal as possible. Definition of minimal could be (1) reducing the number of deleted mappings, or (2) reducing the total amount of deleted mappings’ confidence values. Repair process with new global technique conducted the repair with both minimal definitions. This technique could reduce the number of deleted mappings and total amount of confidence values at the same time. We proposed A * Search method to implement new global technique. This search method was capable to search the shortest path which representing the fewest number of deleted mappings, and also search the cheapest cost which representing the smallest total amount of deleted mappings’ confidence value. A* Search was both complete and optimal to minimize mapping repair size

    Completing and Debugging Ontologies: state of the art and challenges

    Full text link
    As semantically-enabled applications require high-quality ontologies, developing and maintaining ontologies that are as correct and complete as possible is an important although difficult task in ontology engineering. A key step is ontology debugging and completion. In general, there are two steps: detecting defects and repairing defects. In this paper we discuss the state of the art regarding the repairing step. We do this by formalizing the repairing step as an abduction problem and situating the state of the art with respect to this framework. We show that there are still many open research problems and show opportunities for further work and advancing the field.Comment: 56 page

    Building an effective and efficient background knowledge resource to enhance ontology matching

    Get PDF
    International audienceOntology matching is critical for data integration and interoperability. Original ontology matching approaches relied solely on the content of the ontologies to align. However, these approaches are less effective when equivalent concepts have dissimilar labels and are structured with different modeling views. To overcome this semantic heterogeneity, the community has turned to the use of external background knowledge resources. Several methods have been proposed to select ontologies, other than the ones to align, as background knowledge to enhance a given ontology-matching task. However, these methods return a set of complete ontologies, while, in most cases, only fragments of the returned ontologies are effective for discovering new mappings. In this article, we propose an approach to select and build a background knowledge resource with just the right concepts chosen from a set of ontologies, which improves efficiency without loss of effectiveness. The use of background knowledge in ontology matching is a double-edged sword: while it may increase recall (i.e., retrieve more correct mappings), it may lower precision (i.e., produce more incorrect mappings). Therefore, we propose two methods to select the most relevant mappings from the candidate ones: (1)~a selection based on a set of rules and (2)~a selection based on supervised machine learning. Our experiments, conducted on two Ontology Alignment Evaluation Initiative (OAEI) datasets, confirm the effectiveness and efficiency of our approach. Moreover, the F-measure values obtained with our approach are very competitive to those of the state-of-the-art matchers exploiting background knowledge resources

    Semantic Enrichment of Ontology Mappings

    Get PDF
    Schema and ontology matching play an important part in the field of data integration and semantic web. Given two heterogeneous data sources, meta data matching usually constitutes the first step in the data integration workflow, which refers to the analysis and comparison of two input resources like schemas or ontologies. The result is a list of correspondences between the two schemas or ontologies, which is often called mapping or alignment. Many tools and research approaches have been proposed to automatically determine those correspondences. However, most match tools do not provide any information about the relation type that holds between matching concepts, for the simple but important reason that most common match strategies are too simple and heuristic to allow any sophisticated relation type determination. Knowing the specific type holding between two concepts, e.g., whether they are in an equality, subsumption (is-a) or part-of relation, is very important for advanced data integration tasks, such as ontology merging or ontology evolution. It is also very important for mappings in the biological or biomedical domain, where is-a and part-of relations may exceed the number of equality correspondences by far. Such more expressive mappings allow much better integration results and have scarcely been in the focus of research so far. In this doctoral thesis, the determination of the correspondence types in a given mapping is the focus of interest, which is referred to as semantic mapping enrichment. We introduce and present the mapping enrichment tool STROMA, which obtains a pre-calculated schema or ontology mapping and for each correspondence determines a semantic relation type. In contrast to previous approaches, we will strongly focus on linguistic laws and linguistic insights. By and large, linguistics is the key for precise matching and for the determination of relation types. We will introduce various strategies that make use of these linguistic laws and are able to calculate the semantic type between two matching concepts. The observations and insights gained from this research go far beyond the field of mapping enrichment and can be also applied to schema and ontology matching in general. Since generic strategies have certain limits and may not be able to determine the relation type between more complex concepts, like a laptop and a personal computer, background knowledge plays an important role in this research as well. For example, a thesaurus can help to recognize that these two concepts are in an is-a relation. We will show how background knowledge can be effectively used in this instance, how it is possible to draw conclusions even if a concept is not contained in it, how the relation types in complex paths can be resolved and how time complexity can be reduced by a so-called bidirectional search. The developed techniques go far beyond the background knowledge exploitation of previous approaches, and are now part of the semantic repository SemRep, a flexible and extendable system that combines different lexicographic resources. Further on, we will show how additional lexicographic resources can be developed automatically by parsing Wikipedia articles. The proposed Wikipedia relation extraction approach yields some millions of additional relations, which constitute significant additional knowledge for mapping enrichment. The extracted relations were also added to SemRep, which thus became a comprehensive background knowledge resource. To augment the quality of the repository, different techniques were used to discover and delete irrelevant semantic relations. We could show in several experiments that STROMA obtains very good results w.r.t. relation type detection. In a comparative evaluation, it was able to achieve considerably better results than related applications. This corroborates the overall usefulness and strengths of the implemented strategies, which were developed with particular emphasis on the principles and laws of linguistics

    A Language for Inconsistency-Tolerant Ontology Mapping

    Get PDF
    Ontology alignment plays a key role in enabling interoperability among various data sources present in the web. The nature of the world is such, that the same concepts differ in meaning, often so slightly, which makes it difficult to relate these concepts. It is the omni-present heterogeneity that is at the core of the web. The research work presented in this dissertation, is driven by the goal of providing a robust ontology alignment language for the semantic web, as we show that description logics based alignment languages are not suitable for aligning ontologies. The adoption of the semantic web technologies has been consistently on the rise over the past decade, and it continues to show promise. The core component of the semantic web is the set of knowledge representation languages -- mainly the W3C (World Wide Web Consortium) standards Web Ontology Language (OWL), Resource Description Framework (RDF), and Rule Interchange Format (RIF). While these languages have been designed in order to be suitable for the openness and extensibility of the web, they lack certain features which we try to address in this dissertation. One such missing component is the lack of non-monotonic features, in the knowledge representation languages, that enable us to perform common sense reasoning. For example, OWL supports the open world assumption (OWA), which means that knowledge about everything is assumed to be possibly incomplete at any point of time. However, experience has shown that there are situations that require us to assume that certain parts of the knowledge base are complete. Employing the Closed World Assumption (CWA) helps us achieve this. Circumscription is a very well-known approach towards CWA, which provides closed world semantics by employing the idea of minimal models with respect to certain predicates which are closed. We provide the formal semantics of the notion of Grounded Circumscription, which is an extension of circumscription with desirable properties like decidability. We also provide a tableaux calculus to reason over knowledge bases under the notion of grounded circumscription. Another form of common sense logic, is default logic. Default logic provides a way to specify rules that, by default, hold in most cases but not necessarily in all cases. The classic example of such a rule is: If something is a bird then it flies. The power of defaults comes from the ability of the logic to handle exceptions to the default rules. For example, a bird will be assumed to fly by default unless it is an exception, i.e. it belongs to a class of birds that do not fly, like penguins. Interestingly, this property of defaults can be utilized to create mappings between concepts of different ontologies (knowledge bases). We provide a new semantics for the integration of defaults in description logics and show that it improves upon previously known results in literature. In this study, we give various examples to show the utility and advantages of using a default logic based ontology alignment language. We provide the semantics and decidability results of a default based mapping language for tractable fragments of description logics (or OWL). Furthermore, we provide a proof of concept system and qualitative analysis of the results obtained from the system when compared to that of traditional mapping repair techniques

    Evaluating Mapping Repair Systems with Large Biomedical Ontologies

    Full text link
    In this paper we provide empirical evidence of the necessity of integrating mapping repair techniques within the ontology matching process, an aspect that is neglected in many ontology matching systems. We also evaluate the feasibility of using state-of-the-art mapping repair techniques in practice, such as those implemented in Alcomo and LogMap. A preliminary evaluation was conducted in the context of the Ontology Alignment Evaluation Initiative (OAEI) 2012. We extend this evaluation and report about the results in detail
    corecore