409 research outputs found
Ontology mapping: the state of the art
Ontology mapping is seen as a solution provider in today's landscape of ontology research. As the number of ontologies that are made publicly available and accessible on the Web increases steadily, so does the need for applications to use them. A single ontology is no longer enough to support the tasks envisaged by a distributed environment like the Semantic Web. Multiple ontologies need to be accessed from several applications. Mapping could provide a common layer from which several ontologies could be accessed and hence could exchange information in semantically sound manners. Developing such mapping has beeb the focus of a variety of works originating from diverse communities over a number of years. In this article we comprehensively review and present these works. We also provide insights on the pragmatics of ontology mapping and elaborate on a theoretical approach for defining ontology mapping
The Requirements for Ontologies in Medical Data Integration: A Case Study
Evidence-based medicine is critically dependent on three sources of
information: a medical knowledge base, the patients medical record and
knowledge of available resources, including where appropriate, clinical
protocols. Patient data is often scattered in a variety of databases and may,
in a distributed model, be held across several disparate repositories.
Consequently addressing the needs of an evidence-based medicine community
presents issues of biomedical data integration, clinical interpretation and
knowledge management. This paper outlines how the Health-e-Child project has
approached the challenge of requirements specification for (bio-) medical data
integration, from the level of cellular data, through disease to that of
patient and population. The approach is illuminated through the requirements
elicitation and analysis of Juvenile Idiopathic Arthritis (JIA), one of three
diseases being studied in the EC-funded Health-e-Child project.Comment: 6 pages, 1 figure. Presented at the 11th International Database
Engineering & Applications Symposium (Ideas2007). Banff, Canada September
200
Knowledge Components and Methods for Policy Propagation in Data Flows
Data-oriented systems and applications are at the centre of current developments of the World Wide Web (WWW). On the Web of Data (WoD), information sources can be accessed and processed for many purposes. Users need to be aware of any licences or terms of use, which are associated with the data sources they want to use. Conversely, publishers need support in assigning the appropriate policies alongside the data they distribute.
In this work, we tackle the problem of policy propagation in data flows - an expression that refers to the way data is consumed, manipulated and produced within processes. We pose the question of what kind of components are required, and how they can be acquired, managed, and deployed, to support users on deciding what policies propagate to the output of a data-intensive system from the ones associated with its input. We observe three scenarios: applications of the Semantic Web, workflow reuse in Open Science, and the exploitation of urban data in City Data Hubs. Starting from the analysis of Semantic Web applications, we propose a data-centric approach to semantically describe processes as data flows: the Datanode ontology, which comprises a hierarchy of the possible relations between data objects. By means of Policy Propagation Rules, it is possible to link data flow steps and policies derivable from semantic descriptions of data licences. We show how these components can be designed, how they can be effectively managed, and how to reason efficiently with them. In a second phase, the developed components are verified using a Smart City Data Hub as a case study, where we developed an end-to-end solution for policy propagation. Finally, we evaluate our approach and report on a user study aimed at assessing both the quality and the value of the proposed solution
Discovering Implicational Knowledge in Wikidata
Knowledge graphs have recently become the state-of-the-art tool for
representing the diverse and complex knowledge of the world. Examples include
the proprietary knowledge graphs of companies such as Google, Facebook, IBM, or
Microsoft, but also freely available ones such as YAGO, DBpedia, and Wikidata.
A distinguishing feature of Wikidata is that the knowledge is collaboratively
edited and curated. While this greatly enhances the scope of Wikidata, it also
makes it impossible for a single individual to grasp complex connections
between properties or understand the global impact of edits in the graph. We
apply Formal Concept Analysis to efficiently identify comprehensible
implications that are implicitly present in the data. Although the complex
structure of data modelling in Wikidata is not amenable to a direct approach,
we overcome this limitation by extracting contextual representations of parts
of Wikidata in a systematic fashion. We demonstrate the practical feasibility
of our approach through several experiments and show that the results may lead
to the discovery of interesting implicational knowledge. Besides providing a
method for obtaining large real-world data sets for FCA, we sketch potential
applications in offering semantic assistance for editing and curating Wikidata
Ontology Mapping: The State of the Art
Ontology mapping is seen as a solution provider in today\u27s landscape of ontology
research. As the number of ontologies that are made publicly available and
accessible on the Web increases steadily, so does the need for applications to use
them. A single ontology is no longer enough to support the tasks envisaged by a
distributed environment like the Semantic Web. Multiple ontologies need to be
accessed from several applications. Mapping could provide a common layer from which
several ontologies could be accessed and hence could exchange information in
semantically sound manners. Developing such mapping has beeb the focus of a variety
of works originating from diverse communities over a number of years. In this
article we comprehensively review and present these works. We also provide insights
on the pragmatics of ontology mapping and elaborate on a theoretical approach for
defining ontology mapping
Product Family Design Knowledge Representation, Aggregation, Reuse, and Analysis
A flexible information model for systematic development and deployment of product families during all phases of the product realization process is crucial for product-oriented organizations. In current practice, information captured while designing products in a family is often incomplete, unstructured, and is mostly proprietary in nature, making it difficult to index, search, refine, reuse, distribute, browse, aggregate, and analyze knowledge across heterogeneous organizational information systems. To this end, we propose a flexible knowledge management framework to capture, reorganize, and convert both linguistic and parametric product family design information into a unified network, which is called a networked bill of material (NBOM) using formal concept analysis (FCA); encode the NBOM as a cyclic, labeled graph using the Web Ontology Language (OWL) that designers can use to explore, search, and aggregate design information across different phases of product design as well as across multiple products in a product family; and analyze the set of products in a product family based on both linguistic and parametric information. As part of the knowledge management framework, a PostgreSQL database schema has been formulated to serve as a central design repository of product design knowledge, capable of housing the instances of the NBOM. Ontologies encoding the NBOM are utilized as a metalayer in the database schema to connect the design artifacts as part of a graph structure. Representing product families by preconceived common ontologies shows promise in promoting component sharing, and assisting designers search, explore, and analyze linguistic and parametric product family design information. An example involving a family of seven one-time-use cameras with different functions that satisfy a variety of customer needs is presented to demonstrate the implementation of the proposed framework
Learning Description Logic Ontologies: Five Approaches. Where Do They Stand?
Abstract
The quest for acquiring a formal representation of the knowledge of a domain of interest has attracted researchers with various backgrounds into a diverse field called ontology learning. We highlight classical machine learning and data mining approaches that have been proposed for (semi-)automating the creation of description logic (DL) ontologies. These are based on association rule mining, formal concept analysis, inductive logic programming, computational learning theory, and neural networks. We provide an overview of each approach and how it has been adapted for dealing with DL ontologies. Finally, we discuss the benefits and limitations of each of them for learning DL ontologies
- …