87 research outputs found

    A Comparative Study: Change Detection and Querying Dynamic XML Documents

    Get PDF
    The efficient management of the dynamic XML documents is a complex area of research. The changes and size of the XML documents throughout its lifetime are limitless. Change detection is an important part of version management to identify difference between successive versions of a document. Document content is continuously evolving. Users wanted to be able to query previous versions, query changes in documents, as well as to retrieve a particular document version efficiently. In this paper we provide comprehensive comparative analysis of various control schemes for change detection and querying dynamic XML documents

    Temporal and Evolving Data Warehouse Design

    Get PDF

    Report on the 6th ADBIS’2002 conference

    Get PDF
    The 6th East European Conference ADBIS 2002 was held on September~8--11, 2002 in Bratislava, Slovakia. It was organised by the Slovak University of Technology (and, in particular, its Faculty of Electrical Engineering and Information Technology) in Bratislava in co-operation with the ACM SIGMOD, the Moscow ACM SIGMOD Chapter, and Slovak Society for Computer Science. The call for papers attracted 115 submissions from 35~countries. The international program committee, consisting of 43 researchers from 21 countries, selected 25 full papers and 4 short papers for a monograph volume published by the Springer Verlag. Beside those 29 regular papers, the volume includes also 3 invited papers presented at the Conference as invited lectures. Additionally, 20 papers have been selected for the Research communications volume. The authors of accepted papers come from 22~countries of 4 continents, indicating the truly international recognition of the ADBIS conference series. The conference had 104 registered participants from 22~countries and included invited lectures, tutorials, and regular sessions. This report describes the goals of the conference and summarizes the issues discussed during the sessions

    A unified view of data-intensive flows in business intelligence systems : a survey

    Get PDF
    Data-intensive flows are central processes in today’s business intelligence (BI) systems, deploying different technologies to deliver data, from a multitude of data sources, in user-preferred and analysis-ready formats. To meet complex requirements of next generation BI systems, we often need an effective combination of the traditionally batched extract-transform-load (ETL) processes that populate a data warehouse (DW) from integrated data sources, and more real-time and operational data flows that integrate source data at runtime. Both academia and industry thus must have a clear understanding of the foundations of data-intensive flows and the challenges of moving towards next generation BI environments. In this paper we present a survey of today’s research on data-intensive flows and the related fundamental fields of database theory. The study is based on a proposed set of dimensions describing the important challenges of data-intensive flows in the next generation BI setting. As a result of this survey, we envision an architecture of a system for managing the lifecycle of data-intensive flows. The results further provide a comprehensive understanding of data-intensive flows, recognizing challenges that still are to be addressed, and how the current solutions can be applied for addressing these challenges.Peer ReviewedPostprint (author's final draft

    A multidimensional and multiversion structure for OLAP applications

    Full text link
    When changes occur on data organization, conventional multidimensional structures are not adapted because dimensions are supposed to be static. In many cases, especially when time covered by the data warehouse is large, dimensions of the hypercube must be redesigned in order to integrate evolutions. We propose an approach allowing to track history but also to compare data, mapped into static structures. We define a conceptual model building a Mutiversion Fact Table from the Temporal Multidimensional Schema and we introduce the notion of temporal modes of representation corresponding to different ways to analyze data and their evolution

    Modélisation des transformations pour l'évolution de modèles multidimensionnels

    Get PDF
    La modélisation et l'entreposage des données ont constitué, depuis plus d'une décennie, une problématique de recherche pour laquelle différentes approches ont été proposées. Ces approches se focalisent sur des aspects statiques de l'entrepôt de données. Or, l'évolution du système d'information qui alimente un entrepôt peut avoir un impact sur ce dernier et peut conduire, par conséquent, à l'évolution de son modèle multidimensionnel. Dans ce contexte évolutif, nous proposons une démarche dirigée par les modèles pour automatiser la propagation de l'évolution du modèle de la source de données relationnelle vers l'entrepôt. Cette démarche est fondée sur deux modèles d'évolution ainsi qu'un ensemble de règles de transformation formalisées en Query/View/Transformation. Nous développons un prototype logiciel nommé DWE (« Data Warehouse Evolution ») qui supporte cette démarche

    Modélisation des transformations pour l'évolution de modèles multidimensionnels

    Get PDF
    La modélisation et l'entreposage des données ont constitué, depuis plus d'une décennie, une problématique de recherche pour laquelle différentes approches ont été proposées. Ces approches se focalisent sur des aspects statiques de l'entrepôt de données. Or, l'évolution du système d'information qui alimente un entrepôt peut avoir un impact sur ce dernier et peut conduire, par conséquent, à l'évolution de son modèle multidimensionnel. Dans ce contexte évolutif, nous proposons une démarche dirigée par les modèles pour automatiser la propagation de l'évolution du modèle de la source de données relationnelle vers l'entrepôt. Cette démarche est fondée sur deux modèles d'évolution ainsi qu'un ensemble de règles de transformation formalisées en Query/View/Transformation. Nous développons un prototype logiciel nommé DWE (« Data Warehouse Evolution ») qui supporte cette démarche

    Collaborative OLAP with Tag Clouds: Web 2.0 OLAP Formalism and Experimental Evaluation

    Full text link
    Increasingly, business projects are ephemeral. New Business Intelligence tools must support ad-lib data sources and quick perusal. Meanwhile, tag clouds are a popular community-driven visualization technique. Hence, we investigate tag-cloud views with support for OLAP operations such as roll-ups, slices, dices, clustering, and drill-downs. As a case study, we implemented an application where users can upload data and immediately navigate through its ad hoc dimensions. To support social networking, views can be easily shared and embedded in other Web sites. Algorithmically, our tag-cloud views are approximate range top-k queries over spontaneous data cubes. We present experimental evidence that iceberg cuboids provide adequate online approximations. We benchmark several browser-oblivious tag-cloud layout optimizations.Comment: Software at https://github.com/lemire/OLAPTagClou

    A Strategy for Reducing I/O and Improving Query Processing Time in an Oracle Data Warehouse Environment

    Get PDF
    In the current information age as the saying goes, time is money. For the modern information worker, decisions must often be made quickly. Every extra minute spent waiting for critical data could mean the difference between financial gain and financial ruin. Despite the importance of timely data retrieval, many organizations lack even a basic strategy for improving the performance of their data warehouse based reporting systems. This project explores the idea that a strategy making use of three database performance improvement techniques can reduce I/O (input/output operations) and improve query processing time in an information system designed for reporting. To demonstrate that these performance improvement goals can be achieved, queries were run on ordinary tables and then on tables utilizing the performance improvement techniques. The I/O statistics and processing times for the queries were compared to measure the amount of performance improvement. The measurements were also used to explain how these techniques may be more or less effective under certain circumstances, such as when a particular type of query is run. The collected I/O and time based measurements showed a varying degree of improvement for each technique based on the query used. A need to match the types of queries commonly run on the system to the performance improvement technique being implemented was found to be an important consideration. The results indicated that in a reporting environment these performance improvement techniques have the potential to reduce I/O and improve query performance
    • …
    corecore