2,475 research outputs found
Towards an Environment Supporting Resilience, High-Availability, Reproducibility and Reliability for Clud Applications
Cloud Challenge at Utility and Cloud Computing 2015 (UCC 2015), Proceedings of the UCC 2015, Limassol, Cyprus
Supporting Multi-Cloud in Serverless Computing
Serverless computing is a widely adopted cloud execution model composed of
Function-as-a-Service (FaaS) and Backend-as-a-Service (BaaS) offerings. The
increased level of abstraction makes vendor lock-in inherent to serverless
computing, raising more concerns than previous cloud paradigms. Multi-cloud
serverless is a promising emerging approach against vendor lock-in, yet
multiple challenges must be overcome to tap its potential. First, we need to be
aware of both the performance and cost of each FaaS provider. Second, a
multi-cloud architecture must be proposed before deploying a multi-cloud
workflow. Domain-specific serverless offerings must then be integrated into the
multi-cloud architecture to improve performance or save costs. Moreover,
dealing with serverless offerings from multiple providers is challenging.
Finally, we require workload portability support for serverless multi-cloud.
In this paper, we present a multi-cloud library for cross-serverless
offerings. We develop the End Analysis System (EAS) to support comparison among
public FaaS providers in terms of performance and cost. Moreover, we design
proof-of-concept multi-cloud architectures with domain-specific serverless
offerings to alleviate problems such as data gravity. Finally, we deploy
workloads on these architectures to evaluate several public FaaS offerings.Comment: Accepted for the 15th IEEE/ACM International Conference on Utility
and Cloud Computing Companion (UCC'22 Companion
Defeating network jitter for virtual machines
Virtualization based cloud computing hosts networked applications in virtual machines (VMs), and provides each VM the desired degree of performance isolation using resource isolation mechanisms. Existing isolation solutions address heavily on resource proportionality such as CPU, memory and I/O bandwidth, but seldom focus on resource provisioning rate. Even the VM is allocated with adequate resources, if they can not be provided in a timely manner, problems such as network jitter will be very serious and significantly affect the performance of cloud applications like internet audio/video streaming. This paper systematically analyzes and illustrates the causes of unpredictable network latency in virtualized execution environments. We decouple the design goals of resource proportionality from resource provisioning rate, and adopt divide-and-conquer strategy to defeat network jitter for VMs: (1) in VMM CPU scheduling, we differentiate self-initiated I/O from event-triggered I/O, and individually map them to periodic and aperiodic real-time domains to schedule them together; (2) in network traffic shaping of VMs, we introduce the concept of smooth window to smooth network latency and apply closed-loop feedback control to maintain network resource consumption. We implement our solutions in Xen 4.1.0 and Linux 2.6.32.13. The experimental results with both real-life applications and low-level benchmarks show that our solutions can significantly reduce network jitter, and meanwhile effectively maintain resource proportionality.published_or_final_versionThe 4th IEEE International Conference on Utility and Cloud Computing (UCC 2011), Victoria, NSW, 5-8 December 2011. In Proceedings of the 4th IEEE-UCC, 2011, p. 65-7
Performance-oriented Cloud Provisioning: Taxonomy and Survey
Cloud computing is being viewed as the technology of today and the future.
Through this paradigm, the customers gain access to shared computing resources
located in remote data centers that are hosted by cloud providers (CP). This
technology allows for provisioning of various resources such as virtual
machines (VM), physical machines, processors, memory, network, storage and
software as per the needs of customers. Application providers (AP), who are
customers of the CP, deploy applications on the cloud infrastructure and then
these applications are used by the end-users. To meet the fluctuating
application workload demands, dynamic provisioning is essential and this
article provides a detailed literature survey of dynamic provisioning within
cloud systems with focus on application performance. The well-known types of
provisioning and the associated problems are clearly and pictorially explained
and the provisioning terminology is clarified. A very detailed and general
cloud provisioning classification is presented, which views provisioning from
different perspectives, aiding in understanding the process inside-out. Cloud
dynamic provisioning is explained by considering resources, stakeholders,
techniques, technologies, algorithms, problems, goals and more.Comment: 14 pages, 3 figures, 3 table
Autonomic Cloud Computing: Open Challenges and Architectural Elements
As Clouds are complex, large-scale, and heterogeneous distributed systems,
management of their resources is a challenging task. They need automated and
integrated intelligent strategies for provisioning of resources to offer
services that are secure, reliable, and cost-efficient. Hence, effective
management of services becomes fundamental in software platforms that
constitute the fabric of computing Clouds. In this direction, this paper
identifies open issues in autonomic resource provisioning and presents
innovative management techniques for supporting SaaS applications hosted on
Clouds. We present a conceptual architecture and early results evidencing the
benefits of autonomic management of Clouds.Comment: 8 pages, 6 figures, conference keynote pape
Workflow Partitioning and Deployment on the Cloud using Orchestra
Orchestrating service-oriented workflows is typically based on a design model
that routes both data and control through a single point - the centralised
workflow engine. This causes scalability problems that include the unnecessary
consumption of the network bandwidth, high latency in transmitting data between
the services, and performance bottlenecks. These problems are highly prominent
when orchestrating workflows that are composed from services dispersed across
distant geographical locations. This paper presents a novel workflow
partitioning approach, which attempts to improve the scalability of
orchestrating large-scale workflows. It permits the workflow computation to be
moved towards the services providing the data in order to garner optimal
performance results. This is achieved by decomposing the workflow into smaller
sub workflows for parallel execution, and determining the most appropriate
network locations to which these sub workflows are transmitted and subsequently
executed. This paper demonstrates the efficiency of our approach using a set of
experimental workflows that are orchestrated over Amazon EC2 and across several
geographic network regions.Comment: To appear in Proceedings of the IEEE/ACM 7th International Conference
on Utility and Cloud Computing (UCC 2014
SLA-Oriented Resource Provisioning for Cloud Computing: Challenges, Architecture, and Solutions
Cloud computing systems promise to offer subscription-oriented,
enterprise-quality computing services to users worldwide. With the increased
demand for delivering services to a large number of users, they need to offer
differentiated services to users and meet their quality expectations. Existing
resource management systems in data centers are yet to support Service Level
Agreement (SLA)-oriented resource allocation, and thus need to be enhanced to
realize cloud computing and utility computing. In addition, no work has been
done to collectively incorporate customer-driven service management,
computational risk management, and autonomic resource management into a
market-based resource management system to target the rapidly changing
enterprise requirements of Cloud computing. This paper presents vision,
challenges, and architectural elements of SLA-oriented resource management. The
proposed architecture supports integration of marketbased provisioning policies
and virtualisation technologies for flexible allocation of resources to
applications. The performance results obtained from our working prototype
system shows the feasibility and effectiveness of SLA-based resource
provisioning in Clouds.Comment: 10 pages, 7 figures, Conference Keynote Paper: 2011 IEEE
International Conference on Cloud and Service Computing (CSC 2011, IEEE
Press, USA), Hong Kong, China, December 12-14, 201
- …