342 research outputs found
Datacenter Traffic Control: Understanding Techniques and Trade-offs
Datacenters provide cost-effective and flexible access to scalable compute
and storage resources necessary for today's cloud computing needs. A typical
datacenter is made up of thousands of servers connected with a large network
and usually managed by one operator. To provide quality access to the variety
of applications and services hosted on datacenters and maximize performance, it
deems necessary to use datacenter networks effectively and efficiently.
Datacenter traffic is often a mix of several classes with different priorities
and requirements. This includes user-generated interactive traffic, traffic
with deadlines, and long-running traffic. To this end, custom transport
protocols and traffic management techniques have been developed to improve
datacenter network performance.
In this tutorial paper, we review the general architecture of datacenter
networks, various topologies proposed for them, their traffic properties,
general traffic control challenges in datacenters and general traffic control
objectives. The purpose of this paper is to bring out the important
characteristics of traffic control in datacenters and not to survey all
existing solutions (as it is virtually impossible due to massive body of
existing research). We hope to provide readers with a wide range of options and
factors while considering a variety of traffic control mechanisms. We discuss
various characteristics of datacenter traffic control including management
schemes, transmission control, traffic shaping, prioritization, load balancing,
multipathing, and traffic scheduling. Next, we point to several open challenges
as well as new and interesting networking paradigms. At the end of this paper,
we briefly review inter-datacenter networks that connect geographically
dispersed datacenters which have been receiving increasing attention recently
and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial
Quality-of-service management in IP networks
Quality of Service (QoS) in Internet Protocol (IF) Networks has been the subject of
active research over the past two decades. Integrated Services (IntServ) and
Differentiated Services (DiffServ) QoS architectures have emerged as proposed
standards for resource allocation in IF Networks. These two QoS architectures
support the need for multiple traffic queuing systems to allow for resource
partitioning for heterogeneous applications making use of the networks. There have
been a number of specifications or proposals for the number of traffic queuing
classes (Class of Service (CoS)) that will support integrated services in IF Networks,
but none has provided verification in the form of analytical or empirical investigation
to prove that its specification or proposal will be optimum.
Despite the existence of the two standard QoS architectures and the large volume of
research work that has been carried out on IF QoS, its deployment still remains
elusive in the Internet. This is not unconnected with the complexities associated with
some aspects of the standard QoS architectures. [Continues.
Effective Resource and Workload Management in Data Centers
The increasing demand for storage, computation, and business continuity has driven the growth of data centers. Managing data centers efficiently is a difficult task because of the wide variety of datacenter applications, their ever-changing intensities, and the fact that application performance targets may differ widely. Server virtualization has been a game-changing technology for IT, providing the possibility to support multiple virtual machines (VMs) simultaneously. This dissertation focuses on how virtualization technologies can be utilized to develop new tools for maintaining high resource utilization, for achieving high application performance, and for reducing the cost of data center management.;For multi-tiered applications, bursty workload traffic can significantly deteriorate performance. This dissertation proposes an admission control algorithm AWAIT, for handling overloading conditions in multi-tier web services. AWAIT places on hold requests of accepted sessions and refuses to admit new sessions when the system is in a sudden workload surge. to meet the service-level objective, AWAIT serves the requests in the blocking queue with high priority. The size of the queue is dynamically determined according to the workload burstiness.;Many admission control policies are triggered by instantaneous measurements of system resource usage, e.g., CPU utilization. This dissertation first demonstrates that directly measuring virtual machine resource utilizations with standard tools cannot always lead to accurate estimates. A directed factor graph (DFG) model is defined to model the dependencies among multiple types of resources across physical and virtual layers.;Virtualized data centers always enable sharing of resources among hosted applications for achieving high resource utilization. However, it is difficult to satisfy application SLOs on a shared infrastructure, as application workloads patterns change over time. AppRM, an automated management system not only allocates right amount of resources to applications for their performance target but also adjusts to dynamic workloads using an adaptive model.;Server consolidation is one of the key applications of server virtualization. This dissertation proposes a VM consolidation mechanism, first by extending the fair load balancing scheme for multi-dimensional vector scheduling, and then by using a queueing network model to capture the service contentions for a particular virtual machine placement
Recommended from our members
Personal mobile grids with a honeybee inspired resource scheduler
This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The overall aim of the thesis has been to introduce Personal Mobile Grids (PMGrids)
as a novel paradigm in grid computing that scales grid infrastructures to mobile devices and extends grid entities to individual personal users. In this thesis, architectural designs as well as simulation models for PM-Grids are developed.
The core of any grid system is its resource scheduler. However, virtually all current conventional grid schedulers do not address the non-clairvoyant scheduling problem, where job information is not available before the end of execution. Therefore, this thesis proposes a honeybee inspired resource scheduling heuristic for PM-Grids (HoPe) incorporating a radical approach to grid resource scheduling to tackle this problem. A detailed design and implementation of HoPe with a decentralised self-management and adaptive policy are initiated.
Among the other main contributions are a comprehensive taxonomy of grid systems as well as a detailed analysis of the honeybee colony and its nectar acquisition process (NAP), from the resource scheduling perspective, which have not been presented in any previous work, to the best of our knowledge.
PM-Grid designs and HoPe implementation were evaluated thoroughly through a strictly controlled empirical evaluation framework with a well-established heuristic in high throughput computing, the opportunistic scheduling heuristic (OSH), as a benchmark algorithm. Comparisons with optimal values and worst bounds are conducted to gain a clear insight into HoPe behaviour, in terms of stability, throughput, turnaround time and speedup, under different running conditions of number of jobs and grid scales.
Experimental results demonstrate the superiority of HoPe performance where it
has successfully maintained optimum stability and throughput in more than 95%
of the experiments, with HoPe achieving three times better than the OSH under
extremely heavy loads. Regarding the turnaround time and speedup, HoPe has
effectively achieved less than 50% of the turnaround time incurred by the OSH, while doubling its speedup in more than 60% of the experiments.
These results indicate the potential of both PM-Grids and HoPe in realising futuristic grid visions. Therefore considering the deployment of PM-Grids in real life scenarios and the utilisation of HoPe in other parallel processing and high throughput computing systems are recommended
Cost-effective resource management for distributed computing
Current distributed computing and resource management infrastructures (e.g., Cluster and Grid) suffer
from a wide variety of problems related to resource management, which include scalability bottleneck,
resource allocation delay, limited quality-of-service (QoS) support, and lack of cost-aware and service
level agreement (SLA) mechanisms.
This thesis addresses these issues by presenting a cost-effective resource management solution
which introduces the possibility of managing geographically distributed resources in resource units that
are under the control of a Virtual Authority (VA). A VA is a collection of resources controlled, but not
necessarily owned, by a group of users or an authority representing a group of users. It leverages the
fact that different resources in disparate locations will have varying usage levels. By creating smaller
divisions of resources called VAs, users would be given the opportunity to choose between a variety of
cost models, and each VA could rent resources from resource providers when necessary, or could potentially
rent out its own resources when underloaded. The resource management is simplified since the
user and owner of a resource recognize only the VA because all permissions and charges are associated
directly with the VA. The VA is controlled by a ’rental’ policy which is supported by a pool of resources
that the system may rent from external resource providers. As far as scheduling is concerned, the VA is
independent from competitors and can instead concentrate on managing its own resources. As a result,
the VA offers scalable resource management with minimal infrastructure and operating costs.
We demonstrate the feasibility of the VA through both a practical implementation of the prototype
system and an illustration of its quantitative advantages through the use of extensive simulations. First,
the VA concept is demonstrated through a practical implementation of the prototype system. Further, we
perform a cost-benefit analysis of current distributed resource infrastructures to demonstrate the potential
cost benefit of such a VA system. We then propose a costing model for evaluating the cost effectiveness
of the VA approach by using an economic approach that captures revenues generated from applications
and expenses incurred from renting resources. Based on our costing methodology, we present rental
policies that can potentially offer effective mechanisms for running distributed and parallel applications
without a heavy upfront investment and without the cost of maintaining idle resources. By using real
workload trace data, we test the effectiveness of our proposed rental approaches.
Finally, we propose an extension to the VA framework that promotes long-term negotiations and
rentals based on service level agreements or long-term contracts. Based on the extended framework,
we present new SLA-aware policies and evaluate them using real workload traces to demonstrate their effectiveness in improving rental decisions
- …