393 research outputs found
Mapping relational data model to OWL ontology: knowledge conceptualization in OWL
In this paper, we introduce the issues and solutions of using OWL ontology to model extra restriction on 'Properties' of 'Classes' that are not provided by OWL specifications and to represent associations amongst 'Properties' other than 'Classes'. Two specific types of knowledge that cannot be modeled directly using OWL DL elements are identified and presented. Firstly the data value range constraint for a "DatatypeProperty"; secondly the calculation knowledge representation. Our approach to such issues is to conceptualize the knowledge in OWL and map the conceptualization in an implementation. Examples for each type of the knowledge and their OWL code are provided in detail to demonstrate our approach
A Temporal Web Ontology Language
The Web Ontology Language (OWL) is the most expressive standard language for modeling ontologies on the Semantic Web. In this paper, we present a temporal extension of the very expressive fragment SHIN(D) of the OWL-DL language resulting in the tOWL language. Through a layered approach we introduce 3 extensions: i) Concrete Domains, that allows the representation of restrictions using concrete domain binary predicates, ii) Temporal Representation, that introduces timepoints, relations between timepoints, intervals, and Allen’s 13 interval relations into the language, and iii) TimeSlices/Fluents, that implements a perdurantist view on individuals and allows for the representation of complex temporal aspects, such as process state transitions. We illustrate the expressiveness of the newly introduced language by providing a TBox representation of Leveraged Buy Out (LBO) processes in financial applications and an ABox representation of one specific LBO
Towards a Semantic Portal for Oncology using a Description Logic with Fuzzy Concrete Domains
This paper presents three systems that are fully implemented and a proposal for a fourth one. KASIMIR is a knowledge based-system using an ad hoc formalism similar to a simple description logic with concrete domains which is used for representing decision protocols in oncology. FUZZY-KASIMIR is an extension of KASIMIR with fuzzy concrete domains taking into account discontinuities in the decision that are due to numerical thresholds. Another extension of KASIMIR has led to embed it into a semantic portal for oncology, which has been motivated by the need to share knowledge for geographically distributed physicians and has led to change the ad hoc formalism to the standard OWL DL. A combination of these two extensions of KASIMIR is currently under implementation and will lead to a semantic portal for oncology with fuzzy datatypes
Query Rewriting for DL-Lite with n-ary Concrete Domains: Extended Version
We investigate ontology-based query answering (OBQA) in a setting where both the ontology and the query can refer to concrete values such as numbers and strings. In contrast to previous work on this topic, the built-in predicates used to compare values are not restricted to being unary. We introduce restrictions on these predicates and on the ontology language that allow us to reduce OBQA to query answering in databases using the so-called combined rewriting approach. Though at first sight our restrictions are different from the ones used in previous work, we show that our results strictly subsume some of the existing first-order rewritability results for unary predicates.This is an extended version of a paper published in the proceedings of IJCAI 2017
Tractable approximate deduction for OWL
Acknowledgements This work has been partially supported by the European project Marrying Ontologies and Software Technologies (EU ICT2008-216691), the European project Knowledge Driven Data Exploitation (EU FP7/IAPP2011-286348), the UK EPSRC project WhatIf (EP/J014354/1). The authors thank Prof. Ian Horrocks and Dr. Giorgos Stoilos for their helpful discussion on role subsumptions. The authors thank Rafael S. Gonçalves et al. for providing their hotspots ontologies. The authors also thank BoC-group for providing their ADOxx Metamodelling ontologies.Peer reviewedPostprin
Tractable Ontology-Mediated Query Answering with Datatypes
Adding datatypes to ontology-mediated queries (OMQs) often makes query answering hard, even for lightweight languages. As a consequence, the use of datatypes in ontologies, e.g. in OWL 2 QL, has been severely restricted. We propose a new, non-uniform, way of analyzing the data-complexity of OMQ answering with datatypes. Instead of restricting the ontology language we aim at a classification of the patterns of datatype atoms in OMQs into those that can occur in non-tractable OMQs and those that only occur in tractable OMQs. To this end we establish a close link between OMQ answering with datatypes and constraint satisfaction problems (CSPs) over the datatypes. Given that query answering in this setting is undecidable in general already for very simple datatypes, we introduce, borrowing from the database literature, a property of OMQs called the Bounded Match Depth Property (BMDP). We apply the link to CSPs– using results and techniques in universal algebra and model theory–to prove PTIME/co-NP dichotomies for OMQs with the BDMP over Horn-ALCHI extended with (1) all finite datatypes, (2) rational numbers with linear order and (3) certain families of datatypes over the integers with the successor relation
Non classical concept representation and reasoning in formal ontologies
Formal ontologies are nowadays widely considered a standard tool for knowledge
representation and reasoning in the Semantic Web. In this context, they are expected to
play an important role in helping automated processes to access information. Namely:
they are expected to provide a formal structure able to explicate the relationships
between different concepts/terms, thus allowing intelligent agents to interpret, correctly,
the semantics of the web resources improving the performances of the search
technologies.
Here we take into account a problem regarding Knowledge Representation in general,
and ontology based representations in particular; namely: the fact that knowledge
modeling seems to be constrained between conflicting requirements, such as
compositionality, on the one hand and the need to represent prototypical information on
the other. In particular, most common sense concepts seem not to be captured by the
stringent semantics expressed by such formalisms as, for example, Description Logics
(which are the formalisms on which the ontology languages have been built). The aim
of this work is to analyse this problem, suggesting a possible solution suitable for
formal ontologies and semantic web representations.
The questions guiding this research, in fact, have been: is it possible to provide a formal
representational framework which, for the same concept, combines both the classical
modelling view (accounting for compositional information) and defeasible, prototypical
knowledge ? Is it possible to propose a modelling architecture able to provide different
type of reasoning (e.g. classical deductive reasoning for the compositional component
and a non monotonic reasoning for the prototypical one)?
We suggest a possible answer to these questions proposing a modelling framework able
to represent, within the semantic web languages, a multilevel representation of
conceptual information, integrating both classical and non classical (typicality based)
information. Within this framework we hypothesise, at least in principle, the coexistence of multiple reasoning processes involving the different levels of
representation
Automating the multidimensional design of data warehouses
Les experiències prèvies en l'à mbit dels magatzems de dades (o data warehouse), mostren que l'esquema multidimensional del data warehouse ha de ser fruit d'un enfocament hÃbrid; això és, una proposta que consideri tant els requeriments d'usuari com les fonts de dades durant el procés de disseny.Com a qualsevol altre sistema, els requeriments són necessaris per garantir que el sistema desenvolupat satisfà les necessitats de l'usuari. A més, essent aquest un procés de reenginyeria, les fonts de dades s'han de tenir en compte per: (i) garantir que el magatzem de dades resultant pot ésser poblat amb dades de l'organització, i, a més, (ii) descobrir capacitats d'anà lisis no evidents o no conegudes per l'usuari.Actualment, a la literatura s'han presentat diversos mètodes per donar suport al procés de modelatge del magatzem de dades. No obstant això, les propostes basades en un anà lisi dels requeriments assumeixen que aquestos són exhaustius, i no consideren que pot haver-hi informació rellevant amagada a les fonts de dades. Contrà riament, les propostes basades en un anà lisi exhaustiu de les fonts de dades maximitzen aquest enfocament, i proposen tot el coneixement multidimensional que es pot derivar des de les fonts de dades i, conseqüentment, generen massa resultats. En aquest escenari, l'automatització del disseny del magatzem de dades és essencial per evitar que tot el pes de la tasca recaigui en el dissenyador (d'aquesta forma, no hem de confiar únicament en la seva habilitat i coneixement per aplicar el mètode de disseny elegit). A més, l'automatització de la tasca allibera al dissenyador del sempre complex i costós anà lisi de les fonts de dades (que pot arribar a ser inviable per grans fonts de dades).Avui dia, els mètodes automatitzables analitzen en detall les fonts de dades i passen per alt els requeriments. En canvi, els mètodes basats en l'anà lisi dels requeriments no consideren l'automatització del procés, ja que treballen amb requeriments expressats en llenguatges d'alt nivell que un ordenador no pot manegar. Aquesta mateixa situació es dona en els mètodes hÃbrids actual, que proposen un enfocament seqüencial, on l'anà lisi de les dades es complementa amb l'anà lisi dels requeriments, ja que totes dues tasques pateixen els mateixos problemes que els enfocament purs.En aquesta tesi proposem dos mètodes per donar suport a la tasca de modelatge del magatzem de dades: MDBE (Multidimensional Design Based on Examples) and AMDO (Automating the Multidimensional Design from Ontologies). Totes dues consideren els requeriments i les fonts de dades per portar a terme la tasca de modelatge i a més, van ser pensades per superar les limitacions dels enfocaments actuals.1. MDBE segueix un enfocament clà ssic, en el que els requeriments d'usuari són coneguts d'avantmà . Aquest mètode es beneficia del coneixement capturat a les fonts de dades, però guia el procés des dels requeriments i, conseqüentment, és capaç de treballar sobre fonts de dades semà nticament pobres. És a dir, explotant el fet que amb uns requeriments de qualitat, podem superar els inconvenients de disposar de fonts de dades que no capturen apropiadament el nostre domini de treball.2. A diferència d'MDBE, AMDO assumeix un escenari on es disposa de fonts de dades semà nticament riques. Per aquest motiu, dirigeix el procés de modelatge des de les fonts de dades, i empra els requeriments per donar forma i adaptar els resultats generats a les necessitats de l'usuari. En aquest context, a diferència de l'anterior, unes fonts de dades semà nticament riques esmorteeixen el fet de no tenir clars els requeriments d'usuari d'avantmà .Cal notar que els nostres mètodes estableixen un marc de treball combinat que es pot emprar per decidir, donat un escenari concret, quin enfocament és més adient. Per exemple, no es pot seguir el mateix enfocament en un escenari on els requeriments són ben coneguts d'avantmà i en un escenari on aquestos encara no estan clars (un cas recorrent d'aquesta situació és quan l'usuari no té clares les capacitats d'anà lisi del seu propi sistema). De fet, disposar d'uns bons requeriments d'avantmà esmorteeix la necessitat de disposar de fonts de dades semà nticament riques, mentre que a l'inversa, si disposem de fonts de dades que capturen adequadament el nostre domini de treball, els requeriments no són necessaris d'avantmà . Per aquests motius, en aquesta tesi aportem un marc de treball combinat que cobreix tots els possibles escenaris que podem trobar durant la tasca de modelatge del magatzem de dades.Previous experiences in the data warehouse field have shown that the data warehouse multidimensional conceptual schema must be derived from a hybrid approach: i.e., by considering both the end-user requirements and the data sources, as first-class citizens. Like in any other system, requirements guarantee that the system devised meets the end-user necessities. In addition, since the data warehouse design task is a reengineering process, it must consider the underlying data sources of the organization: (i) to guarantee that the data warehouse must be populated from data available within the organization, and (ii) to allow the end-user discover unknown additional analysis capabilities.Currently, several methods for supporting the data warehouse modeling task have been provided. However, they suffer from some significant drawbacks. In short, requirement-driven approaches assume that requirements are exhaustive (and therefore, do not consider the data sources to contain alternative interesting evidences of analysis), whereas data-driven approaches (i.e., those leading the design task from a thorough analysis of the data sources) rely on discovering as much multidimensional knowledge as possible from the data sources. As a consequence, data-driven approaches generate too many results, which mislead the user. Furthermore, the design task automation is essential in this scenario, as it removes the dependency on an expert's ability to properly apply the method chosen, and the need to analyze the data sources, which is a tedious and timeconsuming task (which can be unfeasible when working with large databases). In this sense, current automatable methods follow a data-driven approach, whereas current requirement-driven approaches overlook the process automation, since they tend to work with requirements at a high level of abstraction. Indeed, this scenario is repeated regarding data-driven and requirement-driven stages within current hybrid approaches, which suffer from the same drawbacks than pure data-driven or requirement-driven approaches.In this thesis we introduce two different approaches for automating the multidimensional design of the data warehouse: MDBE (Multidimensional Design Based on Examples) and AMDO (Automating the Multidimensional Design from Ontologies). Both approaches were devised to overcome the limitations from which current approaches suffer. Importantly, our approaches consider opposite initial assumptions, but both consider the end-user requirements and the data sources as first-class citizens.1. MDBE follows a classical approach, in which the end-user requirements are well-known beforehand. This approach benefits from the knowledge captured in the data sources, but guides the design task according to requirements and consequently, it is able to work and handle semantically poorer data sources. In other words, providing high-quality end-user requirements, we can guide the process from the knowledge they contain, and overcome the fact of disposing of bad quality (from a semantical point of view) data sources.2. AMDO, as counterpart, assumes a scenario in which the data sources available are semantically richer. Thus, the approach proposed is guided by a thorough analysis of the data sources, which is properly adapted to shape the output result according to the end-user requirements. In this context, disposing of high-quality data sources, we can overcome the fact of lacking of expressive end-user requirements.Importantly, our methods establish a combined and comprehensive framework that can be used to decide, according to the inputs provided in each scenario, which is the best approach to follow. For example, we cannot follow the same approach in a scenario where the end-user requirements are clear and well-known, and in a scenario in which the end-user requirements are not evident or cannot be easily elicited (e.g., this may happen when the users are not aware of the analysis capabilities of their own sources). Interestingly, the need to dispose of requirements beforehand is smoothed by the fact of having semantically rich data sources. In lack of that, requirements gain relevance to extract the multidimensional knowledge from the sources.So that, we claim to provide two approaches whose combination turns up to be exhaustive with regard to the scenarios discussed in the literaturePostprint (published version
- …