4,930 research outputs found
A Taxonomy of Workflow Management Systems for Grid Computing
With the advent of Grid and application technologies, scientists and
engineers are building more and more complex applications to manage and process
large data sets, and execute scientific experiments on distributed resources.
Such application scenarios require means for composing and executing complex
workflows. Therefore, many efforts have been made towards the development of
workflow management systems for Grid computing. In this paper, we propose a
taxonomy that characterizes and classifies various approaches for building and
executing workflows on Grids. We also survey several representative Grid
workflow systems developed by various projects world-wide to demonstrate the
comprehensiveness of the taxonomy. The taxonomy not only highlights the design
and engineering similarities and differences of state-of-the-art in Grid
workflow systems, but also identifies the areas that need further research.Comment: 29 pages, 15 figure
Dynamic Bandwidth Allocation in Heterogeneous OFDMA-PONs Featuring Intelligent LTE-A Traffic Queuing
This work was supported by the ACCORDANCE project, through the 7th ICT Framework Programme. This is an Accepted Manuscript of an article accepted for publication in Journal of Lightwave Technology following peer review. © 2014 IEEE Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.A heterogeneous, optical/wireless dynamic bandwidth allocation framework is presented, exhibiting intelligent traffic queuing for practically controlling the quality-of-service (QoS) of mobile traffic, backhauled via orthogonal frequency division multiple access–PON (OFDMA-PON) networks. A converged data link layer is presented between long term evolution-advanced (LTE-A) and next-generation passive optical network (NGPON) topologies, extending beyond NGPON2. This is achieved by incorporating in a new protocol design, consistent mapping of LTE-A QCIs and OFDMA-PON queues. Novel inter-ONU algorithms have been developed, based on the distribution of weights to allocate subcarriers to both enhanced node B/optical network units (eNB/ONUs) and residential ONUs, sharing the same infrastructure. A weighted, intra-ONU scheduling mechanism is also introduced to control further the QoS across the network load. The inter and intra-ONU algorithms are both dynamic and adaptive, providing customized solutions to bandwidth allocation for different priority queues at different network traffic loads exhibiting practical fairness in bandwidth distribution. Therefore, middle and low priority packets are not unjustifiably deprived in favor of high priority packets at low network traffic loads. Still the protocol adaptability allows the high priority queues to automatically over perform when the traffic load has increased and the available bandwidth needs to be rationally redistributed. Computer simulations have confirmed that following the application of adaptive weights the fairness index of the new scheme (representing the achieved throughput for each queue), has improved across the traffic load to above 0.9. Packet delay reduction of more than 40ms has been recorded as a result for the low priority queues, while high priories still achieve sufficiently low packet delays in the range of 20 to 30msPeer reviewe
Autonomic Cloud Computing: Open Challenges and Architectural Elements
As Clouds are complex, large-scale, and heterogeneous distributed systems,
management of their resources is a challenging task. They need automated and
integrated intelligent strategies for provisioning of resources to offer
services that are secure, reliable, and cost-efficient. Hence, effective
management of services becomes fundamental in software platforms that
constitute the fabric of computing Clouds. In this direction, this paper
identifies open issues in autonomic resource provisioning and presents
innovative management techniques for supporting SaaS applications hosted on
Clouds. We present a conceptual architecture and early results evidencing the
benefits of autonomic management of Clouds.Comment: 8 pages, 6 figures, conference keynote pape
The Bus Goes Wireless: Routing-Free Data Collection with QoS Guarantees in Sensor Networks
Abstract—We present the low-power wireless bus (LWB), a new communication paradigm for QoS-aware data collection in lowpower sensor networks. The LWB maps all communication onto network floods by using Glossy, an efficient flooding architecture for wireless sensor networks. Therefore, unlike current solutions, the LWB requires no information of the network topology, and inherently supports networks with mobile nodes and multiple data sinks. A LWB prototype implemented in Contiki guarantees bounded end-to-end communication delay and duplicate-free, inorder packet delivery—key QoS requirements in many control and mission-critical applications. Experiments on two testbeds demonstrate that the LWB prototype outperforms state-of-theart data collection and link layer protocols, in terms of reliability and energy efficiency. For instance, we measure an average radio duty cycle of 1.69 % and an overall data yield of 99.97 % in a typical data collection scenario with 85 sensor nodes on Twist. I
A Survey of Techniques For Improving Energy Efficiency in Embedded Computing Systems
Recent technological advances have greatly improved the performance and
features of embedded systems. With the number of just mobile devices now
reaching nearly equal to the population of earth, embedded systems have truly
become ubiquitous. These trends, however, have also made the task of managing
their power consumption extremely challenging. In recent years, several
techniques have been proposed to address this issue. In this paper, we survey
the techniques for managing power consumption of embedded systems. We discuss
the need of power management and provide a classification of the techniques on
several important parameters to highlight their similarities and differences.
This paper is intended to help the researchers and application-developers in
gaining insights into the working of power management techniques and designing
even more efficient high-performance embedded systems of tomorrow
Run-time Spatial Mapping of Streaming Applications to Heterogeneous Multi-Processor Systems
In this paper, we define the problem of spatial mapping. We present reasons why performing spatial mappings at run-time is both necessary and desirable. We propose what is—to our knowledge—the first attempt at a formal description of spatial mappings for the embedded real-time streaming application domain. Thereby, we introduce criteria for a qualitative comparison of these spatial mappings. As an illustration of how our formalization relates to practice, we relate our own spatial mapping algorithm to the formal model
InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services
Cloud computing providers have setup several data centers at different
geographical locations over the Internet in order to optimally serve needs of
their customers around the world. However, existing systems do not support
mechanisms and policies for dynamically coordinating load distribution among
different Cloud-based data centers in order to determine optimal location for
hosting application services to achieve reasonable QoS levels. Further, the
Cloud computing providers are unable to predict geographic distribution of
users consuming their services, hence the load coordination must happen
automatically, and distribution of services must change in response to changes
in the load. To counter this problem, we advocate creation of federated Cloud
computing environment (InterCloud) that facilitates just-in-time,
opportunistic, and scalable provisioning of application services, consistently
achieving QoS targets under variable workload, resource and network conditions.
The overall goal is to create a computing environment that supports dynamic
expansion or contraction of capabilities (VMs, services, storage, and database)
for handling sudden variations in service demands.
This paper presents vision, challenges, and architectural elements of
InterCloud for utility-oriented federation of Cloud computing environments. The
proposed InterCloud environment supports scaling of applications across
multiple vendor clouds. We have validated our approach by conducting a set of
rigorous performance evaluation study using the CloudSim toolkit. The results
demonstrate that federated Cloud computing model has immense potential as it
offers significant performance gains as regards to response time and cost
saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape
- …