60 research outputs found

    Data mapping process to handle semantic data problem on student grading system

    Get PDF
    Many applications are developed on education domain. Information and data for each application are stored in distributed locations with different data representations on each database. This situation leads to heterogeneity at the level of integration data. Heterogeneity data may cause many problems. One major issue is about the semantic relationships data among applications on education domain, in which the learning data may have the same name but with a different meaning, or learning data that has a different name with same meaning. This paper discusses on semantic data mapping process to handle semantic relationships problem on education domain. There are two main parts in the semantic data mapping process. The first part is the semantic data mapping engine to produce data mapping language with turtle (.ttl) file format as a standard XML file schema, that can be used for Local Java Application using Jena Library and Triple Store. The Turtle file contains detail information about data schema of every application inside the database system. The second part is to provide D2R Server that can be accessed from outside environment using HTTP Protocol. This can be done using SPARQL Clients, Linked Data Clients (RDF Formats) and HTML Browser. To implement the semantic data process, this paper focuses on the student grading system in the learning environment of education domain. By following the proposed semantic data mapping process, the turtle file format is produced as a result of the first part of the process. Finally, this file is used to be combined and integrated with other turtle files in order to map and link with other data representation of other applications

    Semantic data mapping technology to solve semantic data problem on heterogeneity aspect

    Get PDF
    The diversity of applications developed with different programming languages, application/data architectures, database systems and representation of data/information leads to heterogeneity issues. One of the problem challenges in the problem of heterogeneity is about heterogeneity data in term of semantic aspect. The semantic aspect is about data that has the same name with different meaning or data that has a different name with the same meaning. The semantic data mapping process is the best solution in the current days to solve semantic data problem. There are many semantic data mapping technologies that have been used in recent years. This research aims to compare and analyze existing semantic data mapping technology using five criteria’s. After comparative and analytical process, this research provides recommendations of appropriate semantic data mapping technology based on several criteria’s. Furthermore, at the end of this research we apply the recommended semantic data mapping technology to be implemented with the real data in the specific application. The result of this research is the semantic data mapping file that contains all data structures in the application data source. This semantic data mapping file can be used to map, share and integrate with other semantic data mapping from other applications and can also be used to integrate with the ontology language

    Anonymous Asserts Profile Protection In Mobile Social Networks

    Get PDF
    In The Current Environment Through Social Network People Can Find Others And Make Their Own Network. An Individual User Can Have Multiple Accounts Of Social Networking Sites. In This Paper We Proposed An Algorithm For Profile Matching In Social Networks, Which Helps To Identify A Particular Person Who Has Multiple Social Networking Accounts And Map His/Her Profile’s Attribute Values With Others In The Same Network To Make A Search Of Friends Easier

    Adaptive Merging on Phase Change Memory

    Full text link
    Indexing is a well-known database technique used to facilitate data access and speed up query processing. Nevertheless, the construction and modification of indexes are very expensive. In traditional approaches, all records in the database table are equally covered by the index. It is not effective, since some records may be queried very often and some never. To avoid this problem, adaptive merging has been introduced. The key idea is to create index adaptively and incrementally as a side-product of query processing. As a result, the database table is indexed partially depending on the query workload. This paper faces a problem of adaptive merging for phase change memory (PCM). The most important features of this memory type are: limited write endurance and high write latency. As a consequence, adaptive merging should be investigated from the scratch. We solve this problem in two steps. First, we apply several PCM optimization techniques to the traditional adaptive merging approach. We prove that the proposed method (eAM) outperforms a traditional approach by 60%. After that, we invent the framework for adaptive merging (PAM) and a new PCM-optimized index. It further improves the system performance by 20% for databases where search queries interleave with data modifications

    Data Mining-based Fragmentation of XML Data Warehouses

    Full text link
    With the multiplication of XML data sources, many XML data warehouse models have been proposed to handle data heterogeneity and complexity in a way relational data warehouses fail to achieve. However, XML-native database systems currently suffer from limited performances, both in terms of manageable data volume and response time. Fragmentation helps address both these issues. Derived horizontal fragmentation is typically used in relational data warehouses and can definitely be adapted to the XML context. However, the number of fragments produced by classical algorithms is difficult to control. In this paper, we propose the use of a k-means-based fragmentation approach that allows to master the number of fragments through its kk parameter. We experimentally compare its efficiency to classical derived horizontal fragmentation algorithms adapted to XML data warehouses and show its superiority
    • …
    corecore