6,463 research outputs found
Energy-Efficient Flow Scheduling and Routing with Hard Deadlines in Data Center Networks
The power consumption of enormous network devices in data centers has emerged
as a big concern to data center operators. Despite many
traffic-engineering-based solutions, very little attention has been paid on
performance-guaranteed energy saving schemes. In this paper, we propose a novel
energy-saving model for data center networks by scheduling and routing
"deadline-constrained flows" where the transmission of every flow has to be
accomplished before a rigorous deadline, being the most critical requirement in
production data center networks. Based on speed scaling and power-down energy
saving strategies for network devices, we aim to explore the most energy
efficient way of scheduling and routing flows on the network, as well as
determining the transmission speed for every flow. We consider two general
versions of the problem. For the version of only flow scheduling where routes
of flows are pre-given, we show that it can be solved polynomially and we
develop an optimal combinatorial algorithm for it. For the version of joint
flow scheduling and routing, we prove that it is strongly NP-hard and cannot
have a Fully Polynomial-Time Approximation Scheme (FPTAS) unless P=NP. Based on
a relaxation and randomized rounding technique, we provide an efficient
approximation algorithm which can guarantee a provable performance ratio with
respect to a polynomial of the total number of flows.Comment: 11 pages, accepted by ICDCS'1
Power Management Techniques for Data Centers: A Survey
With growing use of internet and exponential growth in amount of data to be
stored and processed (known as 'big data'), the size of data centers has
greatly increased. This, however, has resulted in significant increase in the
power consumption of the data centers. For this reason, managing power
consumption of data centers has become essential. In this paper, we highlight
the need of achieving energy efficiency in data centers and survey several
recent architectural techniques designed for power management of data centers.
We also present a classification of these techniques based on their
characteristics. This paper aims to provide insights into the techniques for
improving energy efficiency of data centers and encourage the designers to
invent novel solutions for managing the large power dissipation of data
centers.Comment: Keywords: Data Centers, Power Management, Low-power Design, Energy
Efficiency, Green Computing, DVFS, Server Consolidatio
Software-Defined Cloud Computing: Architectural Elements and Open Challenges
The variety of existing cloud services creates a challenge for service
providers to enforce reasonable Software Level Agreements (SLA) stating the
Quality of Service (QoS) and penalties in case QoS is not achieved. To avoid
such penalties at the same time that the infrastructure operates with minimum
energy and resource wastage, constant monitoring and adaptation of the
infrastructure is needed. We refer to Software-Defined Cloud Computing, or
simply Software-Defined Clouds (SDC), as an approach for automating the process
of optimal cloud configuration by extending virtualization concept to all
resources in a data center. An SDC enables easy reconfiguration and adaptation
of physical resources in a cloud infrastructure, to better accommodate the
demand on QoS through a software that can describe and manage various aspects
comprising the cloud environment. In this paper, we present an architecture for
SDCs on data centers with emphasis on mobile cloud applications. We present an
evaluation, showcasing the potential of SDC in two use cases-QoS-aware
bandwidth allocation and bandwidth-aware, energy-efficient VM placement-and
discuss the research challenges and opportunities in this emerging area.Comment: Keynote Paper, 3rd International Conference on Advances in Computing,
Communications and Informatics (ICACCI 2014), September 24-27, 2014, Delhi,
Indi
Next Generation Cloud Computing: New Trends and Research Directions
The landscape of cloud computing has significantly changed over the last
decade. Not only have more providers and service offerings crowded the space,
but also cloud infrastructure that was traditionally limited to single provider
data centers is now evolving. In this paper, we firstly discuss the changing
cloud infrastructure and consider the use of infrastructure from multiple
providers and the benefit of decentralising computing away from data centers.
These trends have resulted in the need for a variety of new computing
architectures that will be offered by future cloud infrastructure. These
architectures are anticipated to impact areas, such as connecting people and
devices, data-intensive computing, the service space and self-learning systems.
Finally, we lay out a roadmap of challenges that will need to be addressed for
realising the potential of next generation cloud systems.Comment: Accepted to Future Generation Computer Systems, 07 September 201
- …