147,969 research outputs found
Implementing Performance Competitive Logical Recovery
New hardware platforms, e.g. cloud, multi-core, etc., have led to a
reconsideration of database system architecture. Our Deuteronomy project
separates transactional functionality from data management functionality,
enabling a flexible response to exploiting new platforms. This separation
requires, however, that recovery is described logically. In this paper, we
extend current recovery methods to work in this logical setting. While this is
straightforward in principle, performance is an issue. We show how ARIES style
recovery optimizations can work for logical recovery where page information is
not captured on the log. In side-by-side performance experiments using a common
log, we compare logical recovery with a state-of-the art ARIES style recovery
implementation and show that logical redo performance can be competitive.Comment: VLDB201
Query processing on multi-core architectures
The upcoming generation of computer hardware poses several new challenges for database developers and engineers. Software in general and database management systems (DBMSs) in particular will no longer benefit from performance gains of future hardware due to increase clock speed, as it was the case for the last 35 years; instead, the number of cores per CPU will increase steadily. Today’s approach is to run each query on a single core or only a few different cores using parallel query execution. This approach suffers from several problems (e.g. contention problem) and therefore leads to poor speed up and scale up behavior. These observations open several important research questions on how to use the new multi-core CPU architecture for improving the overall performance of DBMSs. This paper outlines our approach for query processing on multi-core CPU architectures. We present an abstract architecture view for multi-core CPUs, meta operators to control and to interact with the hardware, and a new query operator model that makes use of the meta operators to control the parallel execution of a query over different cores. We illustrate how each of these parts fits in our framework for query processing on multi-core architectures
Information integration platform for CIMS
A new information integration platform for computer integrated manufacturing system (CIMS) is presented, which is based on agent and CORBA. CORBA enhances the system integration because it is an industry-standard for interoperable, distributed objects across heterogeneous hardware and software platform. Agent technology is used to improve intelligence of the integration system. In order to implement the information integration platform, we use a network integration server to integrate the network, design a generic database agent to integrate database, adopt multi-agent based architecture to integrate application, and utilize wrapper as a CORBA object to integrate legacy code.published_or_final_versio
Online Learning Algorithm for Time Series Forecasting Suitable for Low Cost Wireless Sensor Networks Nodes
Time series forecasting is an important predictive methodology which can be
applied to a wide range of problems. Particularly, forecasting the indoor
temperature permits an improved utilization of the HVAC (Heating, Ventilating
and Air Conditioning) systems in a home and thus a better energy efficiency.
With such purpose the paper describes how to implement an Artificial Neural
Network (ANN) algorithm in a low cost system-on-chip to develop an autonomous
intelligent wireless sensor network. The present paper uses a Wireless Sensor
Networks (WSN) to monitor and forecast the indoor temperature in a smart home,
based on low resources and cost microcontroller technology as the 8051MCU. An
on-line learning approach, based on Back-Propagation (BP) algorithm for ANNs,
has been developed for real-time time series learning. It performs the model
training with every new data that arrive to the system, without saving enormous
quantities of data to create a historical database as usual, i.e., without
previous knowledge. Consequently to validate the approach a simulation study
through a Bayesian baseline model have been tested in order to compare with a
database of a real application aiming to see the performance and accuracy. The
core of the paper is a new algorithm, based on the BP one, which has been
described in detail, and the challenge was how to implement a computational
demanding algorithm in a simple architecture with very few hardware resources.Comment: 28 pages, Published 21 April 2015 at MDPI's journal "Sensors
Tupleware: Redefining Modern Analytics
There is a fundamental discrepancy between the targeted and actual users of
current analytics frameworks. Most systems are designed for the data and
infrastructure of the Googles and Facebooks of the world---petabytes of data
distributed across large cloud deployments consisting of thousands of cheap
commodity machines. Yet, the vast majority of users operate clusters ranging
from a few to a few dozen nodes, analyze relatively small datasets of up to a
few terabytes, and perform primarily compute-intensive operations. Targeting
these users fundamentally changes the way we should build analytics systems.
This paper describes the design of Tupleware, a new system specifically aimed
at the challenges faced by the typical user. Tupleware's architecture brings
together ideas from the database, compiler, and programming languages
communities to create a powerful end-to-end solution for data analysis. We
propose novel techniques that consider the data, computations, and hardware
together to achieve maximum performance on a case-by-case basis. Our
experimental evaluation quantifies the impact of our novel techniques and shows
orders of magnitude performance improvement over alternative systems
Global state, local decisions: Decentralized NFV for ISPs via enhanced SDN
The network functions virtualization paradigm is rapidly gaining interest among Internet service providers. However, the transition to this paradigm on ISP networks comes with a unique set of challenges: legacy equipment already in place, heterogeneous traffic from multiple clients, and very large scalability requirements. In this article we thoroughly analyze such challenges and discuss NFV design guidelines that address them efficiently. Particularly, we show that a decentralization of NFV control while maintaining global state improves scalability, offers better per-flow decisions and simplifies the implementation of virtual network functions. Building on top of such principles, we propose a partially decentralized NFV architecture enabled via an enhanced software-defined networking infrastructure. We also perform a qualitative analysis of the architecture to identify advantages and challenges. Finally, we determine the bottleneck component, based on the qualitative analysis, which we implement and benchmark in order to assess the feasibility of the architecture.Peer ReviewedPostprint (author's final draft
- …