145,275 research outputs found

    Energy-aware Load Balancing Policies for the Cloud Ecosystem

    Full text link
    The energy consumption of computer and communication systems does not scale linearly with the workload. A system uses a significant amount of energy even when idle or lightly loaded. A widely reported solution to resource management in large data centers is to concentrate the load on a subset of servers and, whenever possible, switch the rest of the servers to one of the possible sleep states. We propose a reformulation of the traditional concept of load balancing aiming to optimize the energy consumption of a large-scale system: {\it distribute the workload evenly to the smallest set of servers operating at an optimal energy level, while observing QoS constraints, such as the response time.} Our model applies to clustered systems; the model also requires that the demand for system resources to increase at a bounded rate in each reallocation interval. In this paper we report the VM migration costs for application scaling.Comment: 10 Page

    SAP: An IoT Application Module Placement Strategy Based on Simulated Annealing Algorithm in Edge-Cloud Computing

    Get PDF
    The Internet of Things (IoT) is rapidly growing and provides the foundation for the development of smart cities, smart home, and health care. With more and more devices connecting to the Internet, huge amounts of data are produced, creating a great challenge for data processing. Traditional cloud computing has the problems of long delays. Edge computing is an extension of cloud computing, processing data at the edge of the network can reduce the long processing delay of cloud computing. Due to the limited computing resources of edge servers, resource management of edge servers has become a critical research problem. However, the structural characteristics of the subtask chain between each pair of sensors and actuators are not considered to address the task scheduling problem in most existing research. To reduce processing latency and energy consumption of the edge-cloud system, we propose a multilayer edge computing system. The application deployed in the system is based on directed digraph. To fully use the edge servers, we proposed an application module placement strategy using Simulated Annealing module Placement (SAP) algorithm. The modules in an application are bounded to each sensor. The SAP algorithm is designed to find a module placement scheme for each sensor and to generate a module chain including the mapping of the module and servers for each sensor. Thus, the edge servers can transmit the tuples in the network with the module chain. To evaluate the efficacy of our algorithm, we simulate the strategy in iFogSim. Results show the scheme is able to achieve significant reductions in latency and energy consumption

    Adaptive runtime techniques for power and resource management on multi-core systems

    Full text link
    Energy-related costs are among the major contributors to the total cost of ownership of data centers and high-performance computing (HPC) clusters. As a result, future data centers must be energy-efficient to meet the continuously increasing computational demand. Constraining the power consumption of the servers is a widely used approach for managing energy costs and complying with power delivery limitations. In tandem, virtualization has become a common practice, as virtualization reduces hardware and power requirements by enabling consolidation of multiple applications on to a smaller set of physical resources. However, administration and management of data center resources have become more complex due to the growing number of virtualized servers installed in data centers. Therefore, designing autonomous and adaptive energy efficiency approaches is crucial to achieve sustainable and cost-efficient operation in data centers. Many modern data centers running enterprise workloads successfully implement energy efficiency approaches today. However, the nature of multi-threaded applications, which are becoming more common in all computing domains, brings additional design and management challenges. Tackling these challenges requires a deeper understanding of the interactions between the applications and the underlying hardware nodes. Although cluster-level management techniques bring significant benefits, node-level techniques provide more visibility into application characteristics, which can then be used to further improve the overall energy efficiency of the data centers. This thesis proposes adaptive runtime power and resource management techniques on multi-core systems. It demonstrates that taking the multi-threaded workload characteristics into account during management significantly improves the energy efficiency of the server nodes, which are the basic building blocks of data centers. The key distinguishing features of this work are as follows: We implement the proposed runtime techniques on state-of-the-art commodity multi-core servers and show that their energy efficiency can be significantly improved by (1) taking multi-threaded application specific characteristics into account while making resource allocation decisions, (2) accurately tracking dynamically changing power constraints by using low-overhead application-aware runtime techniques, and (3) coordinating dynamic adaptive decisions at various layers of the computing stack, specifically at system and application levels. Our results show that efficient resource distribution under power constraints yields energy savings of up to 24% compared to existing approaches, along with the ability to meet power constraints 98% of the time for a diverse set of multi-threaded applications

    The HERA-B database services: for detector configuration, calibration, alignment, slow control and data classification

    Get PDF
    Abstract The database services for the distributed application environment of the HERA-B experiment are presented. Achieving the required 10 6 trigger reduction implies that all reconstruction, including calibration and alignment procedures, must run online, making extensive usage of the database systems. The associations from the events to the database objects are carefully introduced considering efficiency and flexibility. The challenges of managing the slow control information were addressed by introducing data and update objects used in special processing on dedicated servers. The system integrates the DAQ client/server protocols with customized active database servers and relies on a high-performance database support toolkit. For applications that required complex selection mechanisms, as in the data-quality databases, the relevant data is replicated using a relational database management system

    A software architecture for the analysis of large sets of data streams in cloud infrastructures

    Get PDF
    System management algorithms in private andpublic cloud infrastructures have to work with literally thousands of data streams generated from resource, applicationand event monitors. This cloud context opens two novel issuesthat we address in this paper: how to design a softwarearchitecture that is able to gather and analyze all informationwithin real-time constraints; how it is possible to reduce theanalysis of the huge collected data set to the investigationof a reduced set of relevant information. The application ofthe proposed architecture is based on the most advancedsoftware components, and is oriented to the classification of thestatistical behavior of servers and to the analysis of significantstate changes. These results guide model-driven managementsystems to investigate only relevant servers and to applysuitable decision models considering the deter
    corecore