1,610 research outputs found
A Framework for Effective Placement of Virtual Machine Replicas for Highly Available Performance-sensitive Cloud-based Applications
REACTION 2012. 1st International workshop on Real-time and distributed computing in emerging applications. December 4th, 2012, San Juan, Puerto Rico.Applications are increasingly being deployed in
the Cloud due to benefits stemming from economy of scale,
scalability, flexibility and utility-based pricing model. Although
most cloud-based applications have hitherto been enterprisestyle,
there is a new trend towards hosting performancesensitive
applications in the cloud that demand both high
availability and good response times. In the current stateof-
the-art in cloud computing research, there does not exist
solutions that provide both high availability and acceptable
response times to these applications in a way that also optimizes
resource consumption in data centers, which is a key consideration
for cloud providers. This paper addresses this dual
challenge by presenting a design of a fault-tolerant framework
for virtualized data centers that makes two important contributions.
First, it describes an architecture of a fault-tolerance
framework that can be used to automatically deploy replicas
of virtual machines in data centers in a way that optimizes
resources while assures availability and responsiveness. Second,
it describes a specific formulation of a replica deployment
combinatorial optimization problem that can be plugged into
our strategizable deployment framework.This work was supported in part by the National Science
Foundation NSF SHF/CNS Award CNS 0915976 and NSF
CAREER CNS 0845789. Any opinions, findings, and conclusions
or recommendations expressed in this material are
those of the author(s) and do not necessarily reflect the views
of the National Science Foundation
Recommended from our members
Elastic Resource Management in Distributed Clouds
The ubiquitous nature of computing devices and their increasing reliance on remote resources have driven and shaped public cloud platforms into unprecedented large-scale, distributed data centers. Concurrently, a plethora of cloud-based applications are experiencing multi-dimensional workload dynamics---workload volumes that vary along both time and space axes and with higher frequency.
The interplay of diverse workload characteristics and distributed clouds raises several key challenges for efficiently and dynamically managing server resources. First, current cloud platforms impose certain restrictions that might hinder some resource management tasks. Second, an application-agnostic approach might not entail appropriate performance goals, therefore, requires numerous specific methods. Third, provisioning resources outside LAN boundary might incur huge delay which would impact the desired agility.
In this dissertation, I investigate the above challenges and present the design of automated systems that manage resources for various applications in distributed clouds. The intermediate goal of these automated systems is to fully exploit potential benefits such as reduced network latency offered by increasingly distributed server resources. The ultimate goal is to improve end-to-end user response time with novel resource management approaches, within a certain cost budget.
Centered around these two goals, I first investigate how to optimize the location and performance of virtual machines in distributed clouds. I use virtual desktops, mostly serving a single user, as an example use case for developing a black-box approach that ranks virtual machines based on their dynamic latency requirements. Those with high latency sensitivities have a higher priority of being placed or migrated to a cloud location closest to their users. Next, I relax the assumption of well-provisioned virtual machines and look at how to provision enough resources for applications that exhibit both temporal and spatial workload fluctuations. I propose an application-agnostic queueing model that captures the resource utilization and server response time. Building upon this model, I present a geo-elastic provisioning approach---referred as geo-elasticity---for replicable multi-tier applications that can spin up an appropriate amount of server resources in any cloud locations. Last, I explore the benefits of providing geo-elasticity for database clouds, a popular platform for hosting application backends. Performing geo-elastic provisioning for backend database servers entails several challenges that are specific to database workload, and therefore requires tailored solutions. In addition, cloud platforms offer resources at various prices for different locations. Towards this end, I propose a cost-aware geo-elasticity that combines a regression-based workload model and a queueing network capacity model for database clouds.
In summary, hosting a diverse set of applications in an increasingly distributed cloud makes it interesting and necessary to develop new, efficient and dynamic resource management approaches
QoS-aware service continuity in the virtualized edge
5G systems are envisioned to support numerous delay-sensitive applications such as the tactile
Internet, mobile gaming, and augmented reality. Such applications impose new demands on service providers
in terms of the quality of service (QoS) provided to the end-users. Achieving these demands in mobile
5G-enabled networks represent a technical and administrative challenge. One of the solutions proposed is
to provide cloud computing capabilities at the edge of the network. In such vision, services are cloudified
and encapsulated within the virtual machines or containers placed in cloud hosts at the network access layer.
To enable ultrashort processing times and immediate service response, fast instantiation, and migration of
service instances between edge nodes are mandatory to cope with the consequences of user’s mobility.
This paper surveys the techniques proposed for service migration at the edge of the network. We focus
on QoS-aware service instantiation and migration approaches, comparing the mechanisms followed and
emphasizing their advantages and disadvantages. Then, we highlight the open research challenges still left
unhandled.publishe
A multi-criteria decision making approach for scaling and placement of virtual network functions
This paper investigates the joint scaling and placement problem of network services made up of virtual network functions (VNFs) that can be provided inside a cluster managing multiple points of presence (PoPs). Aiming at increasing the VNF service satisfaction rates and minimizing the deployment cost, we use both transport and cloud-aware VNF scaling as well as multi-attribute decision making (MADM) algorithms for VNF placement inside the cluster. The original joint scaling and placement problem is known to be NP-hard and hence the problem is solved by separating scaling and placement problems and solving them individually. The experiments are done using a dataset containing the information of a deployed digital-twin network service. These experiments show that considering transport and cloud parameters during scaling and placement algorithms perform more efficiently than the only cloud based or transport based scaling followed by placement algorithms. One of the MADM algorithms, Total Order Preference by Similarity to the Ideal Solution (TOPSIS), has shown to yield the lowest deployment cost and highest VNF request satisfaction rates compared to only transport or cloud scaling and other investigated MADM algorithms. Our simulation results indicate that considering both transport and cloud parameters in various availability scenarios of cloud and transport resources has significant potential to provide increased request satisfaction rates when VNF scaling and placement using the TOPSIS scheme is performed.This work was partially funded by EC H2020 5GPPP 5Growth Project (Grant 856709), Spanish MINECO Grant TEC2017-88373-R (5G-REFINE), Generalitat de Catalunya Grant 2017 SGR 1195 and the National Program on Equipment and Scientifc and Technical Infrastructure, EQC2018-005257-P under the European Regional Development Fund (FEDER). We would also like to thank Milan Groshev, Carlos Guimarães for providing dataset for scaling of robot manipulator based digital twin service
Computing at massive scale: Scalability and dependability challenges
Large-scale Cloud systems and big data analytics frameworks are now widely used for practical services and applications. However, with the increase of data volume, together with the heterogeneity of workloads and resources, and the dynamic nature of massive user requests, the uncertainties and complexity of resource management and service provisioning increase dramatically, often resulting in poor resource utilization, vulnerable system dependability, and user-perceived performance degradations. In this paper we report our latest understanding of the current and future challenges in this particular area, and discuss both existing and potential solutions to the problems, especially those concerned with system efficiency, scalability and dependability. We first introduce a data-driven analysis methodology for characterizing the resource and workload patterns and tracing performance bottlenecks in a massive-scale distributed computing environment. We then examine and analyze several fundamental challenges and the solutions we are developing to tackle them, including for example incremental but decentralized resource scheduling, incremental messaging communication, rapid system failover, and request handling parallelism. We integrate these solutions with our data analysis methodology in order to establish an engineering approach that facilitates the optimization, tuning and verification of massive-scale distributed systems. We aim to develop and offer innovative methods and mechanisms for future computing platforms that will provide strong support for new big data and IoE (Internet of Everything) applications
- …