10 research outputs found

    Mapping Relational Database to Resource Description Framework using Direct Mapping Method

    Get PDF
    The Resource Description Framework (RDF) is one of the World Wide Web Consortium (W3C) specifications and it is part of W3C semantic web activities. Its main idea is to make the statements about web resources in the form of subject–predicate–object expressions. Relational databases are considered to be the main sources for the web. To integrate them into semantic web, they should be mapped to RDF. The aim of this study is to design a tool that converts relational databases to RDF based on direct mapping method. The designed tool can express the Meta data of relational database in a minimally constrained, flexible, but meaningful way so that web data can be exchanged and integrated without loss of semantics.   &nbsp

    An Algorithm for Mapping Relational Database to Resource Description Framework

    Get PDF
    The Resource Description Framework (RDF) is one of the World Wide Web Consortium (W3C) specifications and it is part of W3C semantic web activities. Its main idea is to make the statements about web resources in the form of subject–predicate–object expressions. Relational databases are considered to be the main sources for the web. To integrate them into semantic web, they should be mapped to RDF. The aim of this study is to design a tool that converts relational databases to RDF based on direct mapping method. The designed tool can express the Meta data of relational database in a minimally constrained, flexible, but meaningful way so that web data can be exchanged and integrated without loss of semantics

    Semantic data mapping technology to solve semantic data problem on heterogeneity aspect

    Get PDF
    The diversity of applications developed with different programming languages, application/data architectures, database systems and representation of data/information leads to heterogeneity issues. One of the problem challenges in the problem of heterogeneity is about heterogeneity data in term of semantic aspect. The semantic aspect is about data that has the same name with different meaning or data that has a different name with the same meaning. The semantic data mapping process is the best solution in the current days to solve semantic data problem. There are many semantic data mapping technologies that have been used in recent years. This research aims to compare and analyze existing semantic data mapping technology using five criteria’s. After comparative and analytical process, this research provides recommendations of appropriate semantic data mapping technology based on several criteria’s. Furthermore, at the end of this research we apply the recommended semantic data mapping technology to be implemented with the real data in the specific application. The result of this research is the semantic data mapping file that contains all data structures in the application data source. This semantic data mapping file can be used to map, share and integrate with other semantic data mapping from other applications and can also be used to integrate with the ontology language

    Data mapping process to handle semantic data problem on student grading system

    Get PDF
    Many applications are developed on education domain. Information and data for each application are stored in distributed locations with different data representations on each database. This situation leads to heterogeneity at the level of integration data. Heterogeneity data may cause many problems. One major issue is about the semantic relationships data among applications on education domain, in which the learning data may have the same name but with a different meaning, or learning data that has a different name with same meaning. This paper discusses on semantic data mapping process to handle semantic relationships problem on education domain. There are two main parts in the semantic data mapping process. The first part is the semantic data mapping engine to produce data mapping language with turtle (.ttl) file format as a standard XML file schema, that can be used for Local Java Application using Jena Library and Triple Store. The Turtle file contains detail information about data schema of every application inside the database system. The second part is to provide D2R Server that can be accessed from outside environment using HTTP Protocol. This can be done using SPARQL Clients, Linked Data Clients (RDF Formats) and HTML Browser. To implement the semantic data process, this paper focuses on the student grading system in the learning environment of education domain. By following the proposed semantic data mapping process, the turtle file format is produced as a result of the first part of the process. Finally, this file is used to be combined and integrated with other turtle files in order to map and link with other data representation of other applications

    RDB to RDF 변환을 위한 의미 정보 보존 맵리듀스 처리

    Get PDF
    학위논문 (석사)-- 서울대학교 대학원 : 전기·컴퓨터공학부, 2015. 8. 김형주.Today, most of the data on the web is stored in relational databases, which is called deep web. Semantic web is a movement to the next generation of the web, where all data are augmented with well-defined semantics and linked together in machine-readable format. RDB2RDF approaches have been proposed and standardized by W3C, which publishes relational data to semantic web by converting relational data into RDF formatted data. We propose a system that automatically transforms relational data into RDF data and creates OWL ontology based on the schema of database. Some approaches have been proposed, but most of them did not fully make use of schema information to extract rich semantics, nor did they experimented on large databases for performance. We utilize Hadoop framework in transformation process, which enables distributed system for scalability. We present mapping rules that implements augmented direct mapping to create local ontology with rich semantics. The results show that our system successfully transforms relational data into RDF data with OWL ontology, with satisfactory performance on large-sized databases.Abstract i Introduction 3 Related Work 7 2.1 Semantic ETL Systems 7 2.2 Hadoop MapReduce 8 2.3 Mapping Approaches 9 Mapping Rules 14 3.1 General Rule 1 19 3.2 General Rule 2 20 3.3 General Rule 3 20 3.4 General Rule 4 21 3.5 General Rule 5 21 3.6 Constraint Rule 1 22 3.7 Constraint Rule 2 22 3.8 Constraint Rule 3 23 3.9 Constraint Rule 4 24 3.10 Constraint Rule 5 24 3.11 Constraint Rule 6 25 3.12 Discussion 26 Our Approach 28 4.1 Preprocessing 28 4.1.1 Schema Caching Method 30 4.1.2 Relational Data 32 4.2 Hadoop Algorithm 33 Experiment 36 5.1 Ontology Extraction 37 5.2 Performance 38 5.3 Scalability 41 Conclusion 42 Reference 44 Appendix 46Maste

    Provenance in Open Data Entity-Centric Aggregation

    Get PDF
    An increasing number of web services these days require combining data from several data providers into an aggregated database. Usually this aggregation is based on the linked data approach. On the other hand, the entity-centric model is a promising data model that outperforms the linked data approach because it solves the lack of explicit semantics and the semantic heterogeneity problems. However, current open data which is available on the web as raw datasets can not be used in the entity-centric model before processing them with an import process to extract the data elements and insert them correctly in the aggregated entity-centric database. It is essential to certify the quality of these imported data elements, especially the background knowledge part which acts as input to semantic computations, because the quality of this part affects directly the quality of the web services which are built on top of it. Furthermore, the aggregation of entities and their attribute values from different sources raises three problems: the need to trace the source of each element, the need to trace the links between entities which can be considered equivalent and the need to handle possible conflicts between different values when they are imported from various data sources. In this thesis, we introduce a new model to certify the quality of a back ground knowledge base which separates linguistic and language independent elements. We also present a pipeline to import entities from open data repositories to add the missing implicit semantics and to eliminate the semantic heterogeneity. Finally, we show how to trace the source of attribute values coming from different data providers; how to choose a strategy for handling possible conflicts between these values; and how to keep the links between identical entities which represent the same real world entity

    OntoIntegra: Método para la integración de datos almacenados en fuentes heterogéneas.

    Get PDF
    Data sources integration is about interconnecting and accessing heterogeneous data sources. With the evolution of the organizations, this topic has become an important research field both for academy and industry. The project “Information and Communication Technologies supporting the educational process and the knowledge management in higher education” (ELINF) needs to integrate heterogeneous data sources for supporting the authority control process in its software. This work aims to develop a method with semantic components to contribute to the integration of authority control related data, required by the ELINF project, which is stored in heterogeneous data sources into a software. The current research describes the developed method. At the same time, it describes the study case used for validating it, through the comparison of a software application developed without the proposed method and another one instantiating the method regarding to its capacity to integrate heterogeneous data sources

    OntoIntegra: Método para la integración de datos almacenados en fuentes heterogéneas.

    Get PDF
    Data sources integration is about interconnecting and accessing heterogeneous data sources. With the evolution of the organizations, this topic has become an important research field both for academy and industry. The project “Information and Communication Technologies supporting the educational process and the knowledge management in higher education” (ELINF) needs to integrate heterogeneous data sources for supporting the authority control process in its software. This work aims to develop a method with semantic components to contribute to the integration of authority control related data, required by the ELINF project, which is stored in heterogeneous data sources into a software. The current research describes the developed method. At the same time, it describes the study case used for validating it, through the comparison of a software application developed without the proposed method and another one instantiating the method regarding to its capacity to integrate heterogeneous data sources

    Contaminant Hydrogeology Knowledge Base (CHKb) of Georgia, USA

    Get PDF
    Hydrogeologists collect data through studies that originate from a diverse and growing set of instruments that measure, for example, geochemical constituents of surface and groundwater. Databases store and publish the collected data on the Web, and the volume of data is quickly increasing, which makes accessing data problematic and time consuming for individuals. One way to overcome this problem is to develop ontology to formally and explicitly represent the domain (e.g., contaminant hydrogeology) knowledge. Using OWL and RDF, contaminant hydrogeology ontology (CHO) is developed to manage hydrological spatial data for Georgia, USA. CHO is a conceptual computer model for the contaminant hydrogeology domain in which concepts (e.g. contaminant, aquifer) and their relationships (e.g. pollutes) are formerly and explicitly defined. Cyberinfrastructure for exposing CHO and datasets (i.e., CHKb) as Linked Data on the Web is developed. Cyberinfrastructure consists of storing, managing, querying, and visualizing CHKb that can be accessed from URL: cho.gsu.edu
    corecore