4,063 research outputs found
InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services
Cloud computing providers have setup several data centers at different
geographical locations over the Internet in order to optimally serve needs of
their customers around the world. However, existing systems do not support
mechanisms and policies for dynamically coordinating load distribution among
different Cloud-based data centers in order to determine optimal location for
hosting application services to achieve reasonable QoS levels. Further, the
Cloud computing providers are unable to predict geographic distribution of
users consuming their services, hence the load coordination must happen
automatically, and distribution of services must change in response to changes
in the load. To counter this problem, we advocate creation of federated Cloud
computing environment (InterCloud) that facilitates just-in-time,
opportunistic, and scalable provisioning of application services, consistently
achieving QoS targets under variable workload, resource and network conditions.
The overall goal is to create a computing environment that supports dynamic
expansion or contraction of capabilities (VMs, services, storage, and database)
for handling sudden variations in service demands.
This paper presents vision, challenges, and architectural elements of
InterCloud for utility-oriented federation of Cloud computing environments. The
proposed InterCloud environment supports scaling of applications across
multiple vendor clouds. We have validated our approach by conducting a set of
rigorous performance evaluation study using the CloudSim toolkit. The results
demonstrate that federated Cloud computing model has immense potential as it
offers significant performance gains as regards to response time and cost
saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape
Software-Defined Cloud Computing: Architectural Elements and Open Challenges
The variety of existing cloud services creates a challenge for service
providers to enforce reasonable Software Level Agreements (SLA) stating the
Quality of Service (QoS) and penalties in case QoS is not achieved. To avoid
such penalties at the same time that the infrastructure operates with minimum
energy and resource wastage, constant monitoring and adaptation of the
infrastructure is needed. We refer to Software-Defined Cloud Computing, or
simply Software-Defined Clouds (SDC), as an approach for automating the process
of optimal cloud configuration by extending virtualization concept to all
resources in a data center. An SDC enables easy reconfiguration and adaptation
of physical resources in a cloud infrastructure, to better accommodate the
demand on QoS through a software that can describe and manage various aspects
comprising the cloud environment. In this paper, we present an architecture for
SDCs on data centers with emphasis on mobile cloud applications. We present an
evaluation, showcasing the potential of SDC in two use cases-QoS-aware
bandwidth allocation and bandwidth-aware, energy-efficient VM placement-and
discuss the research challenges and opportunities in this emerging area.Comment: Keynote Paper, 3rd International Conference on Advances in Computing,
Communications and Informatics (ICACCI 2014), September 24-27, 2014, Delhi,
Indi
Algorithms for advance bandwidth reservation in media production networks
Media production generally requires many geographically distributed actors (e.g., production houses, broadcasters, advertisers) to exchange huge amounts of raw video and audio data. Traditional distribution techniques, such as dedicated point-to-point optical links, are highly inefficient in terms of installation time and cost. To improve efficiency, shared media production networks that connect all involved actors over a large geographical area, are currently being deployed. The traffic in such networks is often predictable, as the timing and bandwidth requirements of data transfers are generally known hours or even days in advance. As such, the use of advance bandwidth reservation (AR) can greatly increase resource utilization and cost efficiency. In this paper, we propose an Integer Linear Programming formulation of the bandwidth scheduling problem, which takes into account the specific characteristics of media production networks, is presented. Two novel optimization algorithms based on this model are thoroughly evaluated and compared by means of in-depth simulation results
Virtual Machines Embedding for Cloud PON AWGR and Server Based Data Centres
In this study, we investigate the embedding of various cloud applications in
PON AWGR and Server Based Data Centres
A Minimum-Cost Flow Model for Workload Optimization on Cloud Infrastructure
Recent technology advancements in the areas of compute, storage and
networking, along with the increased demand for organizations to cut costs
while remaining responsive to increasing service demands have led to the growth
in the adoption of cloud computing services. Cloud services provide the promise
of improved agility, resiliency, scalability and a lowered Total Cost of
Ownership (TCO). This research introduces a framework for minimizing cost and
maximizing resource utilization by using an Integer Linear Programming (ILP)
approach to optimize the assignment of workloads to servers on Amazon Web
Services (AWS) cloud infrastructure. The model is based on the classical
minimum-cost flow model, known as the assignment model.Comment: 2017 IEEE 10th International Conference on Cloud Computin
Energy Efficient Cloud Networks
Cloud computing is expected to be a major factor that will dominate the future Internet service model. This paper summarizes our work on energy efficiency for cloud networks. We develop a framework for studying the energy efficiency of four cloud services in IP over WDM networks: cloud content delivery, storage as a service (StaaS), and virtual machines (VMS) placement for processing applications and infrastructure as a service (IaaS).Our approach is based on the co-optimization of both external network related factors such as whether to geographically centralize or distribute the clouds, the influence of users’ demand distribution, content popularity, access frequency and renewable energy availability and internal capability factors such as the number of servers, switches and routers as well as the amount of storage demanded in each cloud. Our investigation of the different energy efficient approaches is backed with Mixed Integer Linear Programming (MILP) models and real time heuristic
Energy Efficient Resource Allocation for Virtual Network Services with Dynamic Workload in Cloud Data Centers
Title from PDF of title page, viewed on March 21, 2016Dissertation advisor: Baek-Young ChoiVitaIncludes bibliographical references (pages 126-143)Thesis (Ph.D.)--School of Computing and Engineering. University of Missouri--Kansas City, 2016With the rapid proliferation of cloud computing, more and more network services and
applications are deployed on cloud data centers. Their energy consumption and green
house gas emissions have significantly increased. Some efforts have been made to control
and lower energy consumption of data centers such as, proportional energy consuming
hardware, dynamic provisioning, and virtualization machine techniques. However, it is
still common that many servers and network resources are often underutilized, and idle
servers spend a large portion of their peak power consumption.
Network virtualization and resource sharing have been employed to improve energy
efficiency of data centers by aggregating workload to a few physical nodes and switch
the idle nodes to sleep mode. Especially, with the advent of live migration, a virtual node
can be moved from one physical node to another physical node without service disrup
tion. It is possible to save more energy by shrinking virtual nodes to a small set of physical
nodes and turning the idle nodes to sleep mode when the service workload is low, and expanding
virtual nodes to a large set of physical nodes to satisfy QoS requirements when
the service workload is high. When the service provider explicates the desired virtual
network including a specific topology, and a set of virtual nodes with certain resource
demands, the infrastructure provider computes how the given virtual network is embedded to its operated data centers with minimum energy consumption. When the service
provider only gives some description about the network service and the desired QoS requirements, the infrastructure provider has more freedom on how to allocate resources for
the network service.
For the first problem, we consider the evolving workload of the virtual networks
or virtual applications and residual resources in data centers, and build a novel model of
energy efficient virtual network embedding (EE-VNE) in order to minimize energy usage
in the physical network consists of multiple data centers. In this model, both operation
cost for executing network services’ task and migration cost for the live migrations of
virtual nodes are counted toward the total energy consumption. In addition, rather than
random generated physical network topology, we use practical assumption about physical
network topology in our model.
Due to the NP-hardness of the proposed model, we develop a heuristic algorithm for virtual network scheduling and mapping. In doing so, we specifically take the expected
energy consumption at different times, virtual network operation and future migration
costs, and a data center architecture into consideration. Our extensive evaluation results
showthatouralgorithmcouldreduceenergyconsumptionupto40%andtakeuptoa57%
higher number of virtual network requests over other existing virtual mapping schemes.
However, through comparison with CPLEX based exact algorithm, we identify
that there is still a gap between the heuristic solution and the optimal solution. Therefore,
after investigation other solutions, we convert the origin EE-VNE problem to an Ant
Colony Optimization (ACO) problem by building the construction model and presenting
the transition probability formula. Then, ACO based algorithm has been adapted to solve
the ACO-EE-VNE problem. In addition, we reduce the space complexity of ACO-EE
VNE by developing a novel way to track and update the pheromone.
For the second problem, we design a framework to dynamically allocate resources
for a network service by employing container based virtual nodes. In the framework,each
network service would have a pallet container and a set of execution containers. The pal
let container requests resource based on certain strategy, creates execution containers with
assigned resources and manage the life cycle of the containers; while the execution containers execute the assigned job for the network service. Formulations are presented to
optimize resource usage efficiency and save energy consumption for network services
with dynamic workload, and a heuristic algorithm is proposed to solve the optimization
problem. Our numerical results show that container based resource allocation provide
more flexible and saves more cost than virtual service deployment with fixed virtual machines and demands.
In addition, we study the content distribution problem with joint optimization goal
and varied size of contents in cloud storage. Previous research on content distribution
mainly focuses on reducing latency experienced by content customers. A few recent studies address the issue of bandwidth usage in CDNs, as the bandwidth consumption is
an important issue due to its relevance to the cost of content providers. However, few
researches consider both bandwidth consumption and delay performance for the content
providers that use cloud storages with limited budgets, which is the focus of this study. We
develop an efficient light-weight approximation algorithm toward the joint optimization
problem of content placement. We also conduct the analysis of its theoretical complexities. The performance bound of the proposed approximation algorithm exhibits a much
better worst case than those in previous studies. We further extend the approximate algorithm into a distributed version that allows it to promptly react to dynamic changes in
users’ interests. The extensive results from both simulations and Planetlab experiments
exhibit that the performance is near optimal for most of the practical conditions.Introduction -- Related work -- Energy efficient virtual network embedding for green data centers using data center topology and future migration -- Ant colony optimization based energy efficient virtual network embedding -- Energy aware container based resource allocation for virtual services in green data centers -- Achieving optimal content delivery using cloud storage -- Conclusions and future wor
- …