364 research outputs found

    An analysis of the impact of datacenter temperature on energy efficiency

    Get PDF
    Thesis (S.M. in Engineering and Management)--Massachusetts Institute of Technology, Engineering Systems Division, System Design and Management Program; in conjunction with the SDM Fellows Program, 2012.Cataloged from PDF version of thesis.Includes bibliographical references (p. 65-66).The optimal air temperature for datacenters is one of ways to improve energy efficiency of datacenter cooling systems. Many datacenter owners have been interested in raising the room temperature as a quick and simple method to increase energy efficiency. The purpose of this paper is both to provide recommendations on maximizing the energy efficiency of datacenters by optimizing datacenter temperature setpoint, and to understand the drivers of datacenter costs. This optimization and the potential energy savings used in cooling system can drive higher energy use in IT equipment and may not be a good trade off. For this reason, this paper provided a detailed look at the overall effect on energy of temperature changes in order to figure out the optimal datacenter temperature setpoint. Since this optimal temperature range varies by equipment and other factors in the datacenter, each datacenter should identify its appropriate temperature based on the optimization calculation in this paper. Sensitivity analysis is used to identify the drivers of the cost of ownership in a datacenter and to identify opportunities for datacenter efficiency improvement. The model is also used to evaluate potential datacenter efficiency.by Heechang Lee.S.M.in Engineering and Managemen

    Large-scale End-of-Life Prediction of Hard Disks in Distributed Datacenters

    Full text link
    On a daily basis, data centers process huge volumes of data backed by the proliferation of inexpensive hard disks. Data stored in these disks serve a range of critical functional needs from financial, and healthcare to aerospace. As such, premature disk failure and consequent loss of data can be catastrophic. To mitigate the risk of failures, cloud storage providers perform condition-based monitoring and replace hard disks before they fail. By estimating the remaining useful life of hard disk drives, one can predict the time-to-failure of a particular device and replace it at the right time, ensuring maximum utilization whilst reducing operational costs. In this work, large-scale predictive analyses are performed using severely skewed health statistics data by incorporating customized feature engineering and a suite of sequence learners. Past work suggests using LSTMs as an excellent approach to predicting remaining useful life. To this end, we present an encoder-decoder LSTM model where the context gained from understanding health statistics sequences aid in predicting an output sequence of the number of days remaining before a disk potentially fails. The models developed in this work are trained and tested across an exhaustive set of all of the 10 years of S.M.A.R.T. health data in circulation from Backblaze and on a wide variety of disk instances. It closes the knowledge gap on what full-scale training achieves on thousands of devices and advances the state-of-the-art by providing tangible metrics for evaluation and generalization for practitioners looking to extend their workflow to all years of health data in circulation across disk manufacturers. The encoder-decoder LSTM posted an RMSE of 0.83 during training and 0.86 during testing over the exhaustive 10 year data while being able to generalize competitively over other drives from the Seagate family.Comment: 8 pages, 9 figures and 6 table

    Is Hot IT a False Economy? An Analysis of Server and Data Center Energy Efficiency as Temperatures Rise

    Get PDF
    As demand for digital services grows, there is need to improve efficiency and reduce the environmental impact of data centers. The largest energy consumer in any data center is the IT, followed by the systems dedicated to cooling. Aiming to improve efficiency, and driven by metrics like PUE, there is a trend towards running data centers hotter to reduce the cooling energy. There is little research investigating the effect this will have on the IT beyond failure rates. To ensure overall efficiency is improving, we must view the data center as a system of systems, taking a holistic view rather than focusing on individual sub-systems. In this paper we use industry standard benchmarks and a wind-tunnel to profile typical enterprise IT. We analyze the effect of environmental conditions on IT efficiency, showing minor increases in temperature or pressure impact the efficiency of servers. Using an idealized, simulated data center case study we show that the interaction between cooling systems, server behaviour and local climate are non-trivial and increasing temperatures has potential to worsen efficiency

    Cloud Computing cost and energy optimization through Federated Cloud SoS

    Get PDF
    2017 Fall.Includes bibliographical references.The two most significant differentiators amongst contemporary Cloud Computing service providers have increased green energy use and datacenter resource utilization. This work addresses these two issues from a system's architectural optimization viewpoint. The proposed approach herein, allows multiple cloud providers to utilize their individual computing resources in three ways by: (1) cutting the number of datacenters needed, (2) scheduling available datacenter grid energy via aggregators to reduce costs and power outages, and lastly by (3) utilizing, where appropriate, more renewable and carbon-free energy sources. Altogether our proposed approach creates an alternative paradigm for a Federated Cloud SoS approach. The proposed paradigm employs a novel control methodology that is tuned to obtain both financial and environmental advantages. It also supports dynamic expansion and contraction of computing capabilities for handling sudden variations in service demand as well as for maximizing usage of time varying green energy supplies. Herein we analyze the core SoS requirements, concept synthesis, and functional architecture with an eye on avoiding inadvertent cascading conditions. We suggest a physical architecture that diminishes unwanted outcomes while encouraging desirable results. Finally, in our approach, the constituent cloud services retain their independent ownership, objectives, funding, and sustainability means. This work analyzes the core SoS requirements, concept synthesis, and functional architecture. It suggests a physical structure that simulates the primary SoS emergent behavior to diminish unwanted outcomes while encouraging desirable results. The report will analyze optimal computing generation methods, optimal energy utilization for computing generation as well as a procedure for building optimal datacenters using a unique hardware computing system design based on the openCompute community as an illustrative collaboration platform. Finally, the research concludes with security features cloud federation requires to support to protect its constituents, its constituents tenants and itself from security risks

    A unified model for holistic power usage in cloud datacenter servers

    Get PDF
    Cloud datacenters are compute facilities formed by hundreds and thousands of heterogeneous servers requiring significant power requirements to operate effectively. Servers are composed by multiple interacting sub-systems including applications, microelectronic processors, and cooling which reflect their respective power profiles via different parameters. What is presently unknown is how to accurately model the holistic power usage of the entire server when including all these sub-systems together. This becomes increasingly challenging when considering diverse utilization patterns, server hardware characteristics, air and liquid cooling techniques, and importantly quantifying the non-electrical energy cost imposed by cooling operation. Such a challenge arises due to the need for multi-disciplinary expertise required to study server operation holistically. This work provides a unified model for capturing holistic power usage within Cloud datacenter servers. Constructed through controlled laboratory experiments, the model captures the relationship of server power usage between software, hardware, and cooling agnostic of architecture and cooling type (air and liquid). An exciting prospect is the ability to quantify the amount of non-electrical power consumed through cooling, allowing for more realistic and accurate server power profiles. This work represents the first empirically supported analysis and modeling of holistic power usage for Cloud datacenter servers, and bridges a significant gap between computer science and mechanical engineering research. Model validation through experiments demonstrates an average standard error of 3% for server power usage within both air and liquid cooled environments
    • …
    corecore