4,930 research outputs found

    A Taxonomy of Workflow Management Systems for Grid Computing

    Full text link
    With the advent of Grid and application technologies, scientists and engineers are building more and more complex applications to manage and process large data sets, and execute scientific experiments on distributed resources. Such application scenarios require means for composing and executing complex workflows. Therefore, many efforts have been made towards the development of workflow management systems for Grid computing. In this paper, we propose a taxonomy that characterizes and classifies various approaches for building and executing workflows on Grids. We also survey several representative Grid workflow systems developed by various projects world-wide to demonstrate the comprehensiveness of the taxonomy. The taxonomy not only highlights the design and engineering similarities and differences of state-of-the-art in Grid workflow systems, but also identifies the areas that need further research.Comment: 29 pages, 15 figure

    Dynamic Bandwidth Allocation in Heterogeneous OFDMA-PONs Featuring Intelligent LTE-A Traffic Queuing

    Get PDF
    This work was supported by the ACCORDANCE project, through the 7th ICT Framework Programme. This is an Accepted Manuscript of an article accepted for publication in Journal of Lightwave Technology following peer review. © 2014 IEEE Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.A heterogeneous, optical/wireless dynamic bandwidth allocation framework is presented, exhibiting intelligent traffic queuing for practically controlling the quality-of-service (QoS) of mobile traffic, backhauled via orthogonal frequency division multiple access–PON (OFDMA-PON) networks. A converged data link layer is presented between long term evolution-advanced (LTE-A) and next-generation passive optical network (NGPON) topologies, extending beyond NGPON2. This is achieved by incorporating in a new protocol design, consistent mapping of LTE-A QCIs and OFDMA-PON queues. Novel inter-ONU algorithms have been developed, based on the distribution of weights to allocate subcarriers to both enhanced node B/optical network units (eNB/ONUs) and residential ONUs, sharing the same infrastructure. A weighted, intra-ONU scheduling mechanism is also introduced to control further the QoS across the network load. The inter and intra-ONU algorithms are both dynamic and adaptive, providing customized solutions to bandwidth allocation for different priority queues at different network traffic loads exhibiting practical fairness in bandwidth distribution. Therefore, middle and low priority packets are not unjustifiably deprived in favor of high priority packets at low network traffic loads. Still the protocol adaptability allows the high priority queues to automatically over perform when the traffic load has increased and the available bandwidth needs to be rationally redistributed. Computer simulations have confirmed that following the application of adaptive weights the fairness index of the new scheme (representing the achieved throughput for each queue), has improved across the traffic load to above 0.9. Packet delay reduction of more than 40ms has been recorded as a result for the low priority queues, while high priories still achieve sufficiently low packet delays in the range of 20 to 30msPeer reviewe

    Autonomic Cloud Computing: Open Challenges and Architectural Elements

    Full text link
    As Clouds are complex, large-scale, and heterogeneous distributed systems, management of their resources is a challenging task. They need automated and integrated intelligent strategies for provisioning of resources to offer services that are secure, reliable, and cost-efficient. Hence, effective management of services becomes fundamental in software platforms that constitute the fabric of computing Clouds. In this direction, this paper identifies open issues in autonomic resource provisioning and presents innovative management techniques for supporting SaaS applications hosted on Clouds. We present a conceptual architecture and early results evidencing the benefits of autonomic management of Clouds.Comment: 8 pages, 6 figures, conference keynote pape

    The Bus Goes Wireless: Routing-Free Data Collection with QoS Guarantees in Sensor Networks

    Get PDF
    Abstract—We present the low-power wireless bus (LWB), a new communication paradigm for QoS-aware data collection in lowpower sensor networks. The LWB maps all communication onto network floods by using Glossy, an efficient flooding architecture for wireless sensor networks. Therefore, unlike current solutions, the LWB requires no information of the network topology, and inherently supports networks with mobile nodes and multiple data sinks. A LWB prototype implemented in Contiki guarantees bounded end-to-end communication delay and duplicate-free, inorder packet delivery—key QoS requirements in many control and mission-critical applications. Experiments on two testbeds demonstrate that the LWB prototype outperforms state-of-theart data collection and link layer protocols, in terms of reliability and energy efficiency. For instance, we measure an average radio duty cycle of 1.69 % and an overall data yield of 99.97 % in a typical data collection scenario with 85 sensor nodes on Twist. I

    A Survey of Techniques For Improving Energy Efficiency in Embedded Computing Systems

    Full text link
    Recent technological advances have greatly improved the performance and features of embedded systems. With the number of just mobile devices now reaching nearly equal to the population of earth, embedded systems have truly become ubiquitous. These trends, however, have also made the task of managing their power consumption extremely challenging. In recent years, several techniques have been proposed to address this issue. In this paper, we survey the techniques for managing power consumption of embedded systems. We discuss the need of power management and provide a classification of the techniques on several important parameters to highlight their similarities and differences. This paper is intended to help the researchers and application-developers in gaining insights into the working of power management techniques and designing even more efficient high-performance embedded systems of tomorrow

    Run-time Spatial Mapping of Streaming Applications to Heterogeneous Multi-Processor Systems

    Get PDF
    In this paper, we define the problem of spatial mapping. We present reasons why performing spatial mappings at run-time is both necessary and desirable. We propose what is—to our knowledge—the first attempt at a formal description of spatial mappings for the embedded real-time streaming application domain. Thereby, we introduce criteria for a qualitative comparison of these spatial mappings. As an illustration of how our formalization relates to practice, we relate our own spatial mapping algorithm to the formal model

    InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services

    Full text link
    Cloud computing providers have setup several data centers at different geographical locations over the Internet in order to optimally serve needs of their customers around the world. However, existing systems do not support mechanisms and policies for dynamically coordinating load distribution among different Cloud-based data centers in order to determine optimal location for hosting application services to achieve reasonable QoS levels. Further, the Cloud computing providers are unable to predict geographic distribution of users consuming their services, hence the load coordination must happen automatically, and distribution of services must change in response to changes in the load. To counter this problem, we advocate creation of federated Cloud computing environment (InterCloud) that facilitates just-in-time, opportunistic, and scalable provisioning of application services, consistently achieving QoS targets under variable workload, resource and network conditions. The overall goal is to create a computing environment that supports dynamic expansion or contraction of capabilities (VMs, services, storage, and database) for handling sudden variations in service demands. This paper presents vision, challenges, and architectural elements of InterCloud for utility-oriented federation of Cloud computing environments. The proposed InterCloud environment supports scaling of applications across multiple vendor clouds. We have validated our approach by conducting a set of rigorous performance evaluation study using the CloudSim toolkit. The results demonstrate that federated Cloud computing model has immense potential as it offers significant performance gains as regards to response time and cost saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape
    corecore