158 research outputs found
Implementation of a canonical native storage for XML
The Extensible Markup Language (XML) is a simple, natural, but powerful language to describe data and metada and it is being used widely. However, memory limitation becomes the main problem when XML queries are executed against those large XML documents. CanStoreX is a solution for this. This thesis describes the overall design, architecture and implementation of the CanStoreX XML data management system. CanStoreX partitions a large XML document into pages by adding storage-facilitating nodes. As a native XML DBMS, CanStoreX consists of four layers: disk space management, buffer management, CanStoreX DOM API, and XPath query engine. In order to load XML documents into our CanStoreX system, a dynamic bottom-up loading algorithm is proposed. The experimental results show that the CanStoreX system can handle XML data that are up to two orders of magnitude larger than what is currently possible. Some hints are predicted in here to make the CanStoreX storage highly scalable to handle terabyte data
Integrating data warehouses with web data : a survey
This paper surveys the most relevant research on combining Data Warehouse (DW) and Web data. It studies the XML
technologies that are currently being used to integrate, store, query, and retrieve Web data and their application to DWs. The paper
reviews different DW distributed architectures and the use of XML languages as an integration tool in these systems. It also introduces
the problem of dealing with semistructured data in a DW. It studies Web data repositories, the design of multidimensional databases for
XML data sources, and the XML extensions of OnLine Analytical Processing techniques. The paper addresses the application of
information retrieval technology in a DW to exploit text-rich document collections. The authors hope that the paper will help to discover
the main limitations and opportunities that offer the combination of the DW and the Web fields, as well as to identify open research
line
Anatomy of a Native XML Base Management System
Several alternatives to manage large XML document collections exist, ranging from file systems over relational or other database systems to specifically tailored XML repositories. In this paper we give a tour of Natix, a database management system designed from scratch for storing and processing XML data. Contrary to the common belief that management of XML data is just another application for traditional databases like relational systems, we illustrate how almost every component in a database system is affected in terms of adequacy and performance. We show how to design and optimize areas such as storage, transaction management comprising recovery and multi-user synchronisation as well as query processing for XML
Data in Business Process Models. A Preliminary Empirical Study
Traditional activity-centric process modeling languages treat data as simple black boxes acting as input or output for activities. Many alternate and emerging process modeling paradigms, such as case handling and artifact-centric process modeling, give data a more central role. This is achieved by introducing lifecycles and states for data objects, which is beneficial when modeling data-or knowledge-intensive processes. We assume that traditional activity-centric process modeling languages lack the capabilities to adequately capture the complexity of such processes. To verify this assumption we conducted an online interview among BPM experts. The results not only allow us to identify various profiles of persons modeling business processes, but also the problems that exist in contemporary modeling languages w.r.t. The modeling of business data. Overall, this preliminary empirical study confirms the necessity of data-awareness in process modeling notations in general
Leveraging Naval Riverine forces to achieve information superiority in stability operations
The conflicts of Iraq and Afghanistan have provided an undeniable storyline: U.S. forces can conduct a conventional mission better than any in the world, but that mission, accomplished in short order, leaves behind a situation for which conventional forces and equipment are ill-prepared. This situation requires a new mission: Stability Operations. The blue-water is not where these 21st century conflicts will likely take place, and forces such as the U.S. Navy Riverines are among the many forces that provide a capability to integrate and communicate with local populations that cannot be matched by blue-water forces. While the riverine force's mission set is one that could become heavily utilized in stability operations, the ability to conduct those missions is currently hindered by a lack of implementation of information technology. The current disadvantages that greatly increase operational risk include a reduced capability to engage the population, reduced situational awareness, and limited communication reach-back capability. A riverine force properly equipped with and trained with biometric, unmanned, and information sharing systems would provide the NECC, and U.S. Navy as a whole, a more comprehensive ability to conduct stability operations in brown-water areas, something no other conventional Navy unit can currently accomplish.http://archive.org/details/leveragingnavalr109455075US Navy (USN) authorApproved for public release; distribution is unlimited
Assessment of security extended XML-based Management
The emergence of new management paradigms having XML as a core foundation block demands a comprehensive analysis of their security and performance issues. This paper presents an extension to the existing NetConf protocol. This extension consists of a security architecture and some advanced XML specific features. We describe a series of experiments addressing the performance and operational aspects of our developed implementation and provide grounded answers to issues of significant relevancy to the research community
Recommended from our members
Customized Infrastructures for Monitoring Business Processes
Process enactment technology provides native tools and add-ons for monitoring, such as APIs and monitoring consoles, which are usually highly entangled with the underlying process enactment logic and not customizable by process users. In such a case, all users access the same set of monitoring data and functions and process management resources may be allocated for monitoring concerns not of interest for users. In this context, we present a model and a tool for customized process monitoring infrastructures executing on top of existing process enactment technology. The model classifies the options about monitoring over which the preferences of process users may diverge. The tool implements the proposed model, generating customized process monitoring infrastructures embedding the business logic of the monitoring options chosen by process users
- …