635 research outputs found

    An Algorithm for Network and Data-aware Placement of Multi-Tier Applications in Cloud Data Centers

    Full text link
    Today's Cloud applications are dominated by composite applications comprising multiple computing and data components with strong communication correlations among them. Although Cloud providers are deploying large number of computing and storage devices to address the ever increasing demand for computing and storage resources, network resource demands are emerging as one of the key areas of performance bottleneck. This paper addresses network-aware placement of virtual components (computing and data) of multi-tier applications in data centers and formally defines the placement as an optimization problem. The simultaneous placement of Virtual Machines and data blocks aims at reducing the network overhead of the data center network infrastructure. A greedy heuristic is proposed for the on-demand application components placement that localizes network traffic in the data center interconnect. Such optimization helps reducing communication overhead in upper layer network switches that will eventually reduce the overall traffic volume across the data center. This, in turn, will help reducing packet transmission delay, increasing network performance, and minimizing the energy consumption of network components. Experimental results demonstrate performance superiority of the proposed algorithm over other approaches where it outperforms the state-of-the-art network-aware application placement algorithm across all performance metrics by reducing the average network cost up to 67% and network usage at core switches up to 84%, as well as increasing the average number of application deployments up to 18%.Comment: Submitted for publication consideration for the Journal of Network and Computer Applications (JNCA). Total page: 28. Number of figures: 15 figure

    Energy-aware Load Balancing Policies for the Cloud Ecosystem

    Full text link
    The energy consumption of computer and communication systems does not scale linearly with the workload. A system uses a significant amount of energy even when idle or lightly loaded. A widely reported solution to resource management in large data centers is to concentrate the load on a subset of servers and, whenever possible, switch the rest of the servers to one of the possible sleep states. We propose a reformulation of the traditional concept of load balancing aiming to optimize the energy consumption of a large-scale system: {\it distribute the workload evenly to the smallest set of servers operating at an optimal energy level, while observing QoS constraints, such as the response time.} Our model applies to clustered systems; the model also requires that the demand for system resources to increase at a bounded rate in each reallocation interval. In this paper we report the VM migration costs for application scaling.Comment: 10 Page

    Dynamic Scaling for Service Oriented Applications: Implications of Virtual Machine Placement on IaaS Clouds

    Get PDF
    Abstraction of physical hardware using infrastructure-as-a-service (IaaS) clouds leads to the simplistic view that resources are homogeneous and that infinite scaling is possible with linear increases in performance. Support for autonomic scaling of multi-tier service oriented applications requires determination of when, what, and where to scale. \u27When\u27 is addressed by hotspot detection schemes using techniques including performance modeling and time series analysis. \u27What\u27 relates to determining the quantity and size of new resources to provision. \u27Where\u27 involves identification of the best location(s) to provision new resources. In this paper we investigate primarily \u27where\u27 new infrastructure should be provisioned, and secondly \u27what\u27 the infrastructure should be. Dynamic scaling of infrastructure for service oriented applications requires rapid response to changes in demand to meet application quality-of-service requirements. We investigate the performance and resource cost implications of VM placement when dynamically scaling server infrastructure of service oriented applications . We evaluate dynamic scaling in the context of providing modeling-as-a-service for two environmental science models

    A Systematic Mapping Study of Empirical Studies on Software Cloud Testing Methods

    Get PDF
    Context: Software has become more complicated, dynamic, and asynchronous than ever, making testing more challenging. With the increasing interest in the development of cloud computing, and increasing demand for cloud-based services, it has become essential to systematically review the research in the area of software testing in the context of cloud environments. Objective: The purpose of this systematic mapping study is to provide an overview of the empirical research in the area of software cloud-based testing, in order to build a classification scheme. We investigate functional and non-functional testing methods, the application of these methods, and the purpose of testing using these methods. Method: We searched for electronically available papers in order to find relevant literature and to extract and analyze data about the methods used. Result: We identified 69 primary studies reported in 75 research papers published in academic journals, conferences, and edited books. Conclusion: We found that only a minority of the studies combine rigorous statistical analysis with quantitative results. The majority of the considered studies present early results, using a single experiment to evaluate their proposed solution

    Cloud migration patterns: a multi-cloud service architecture perspective

    Get PDF
    Many organizations migrate their on-premise software systems to the cloud. However, current coarse-grained cloud migration solutions have made a transparent migration of on-premise applications to the cloud a difficult, sometimes trial-and-error based endeavor. This paper suggests a catalogue of fine-grained service-based cloud architecture migration patterns that target multi-cloud settings and are specified with architectural notations. The proposed migration patterns are based on empirical evi-dence from a number of migration projects, best practices for cloud architectures and a systematic literature review of existing research. The pattern catalogue allows an or-ganization to (1) select appropriate architecture migration patterns based on their ob-jectives, (2) compose them to define a migration plan, and (3) extend them based on the identification of new patterns in new contexts

    Pattern-based multi-cloud architecture migration

    Get PDF
    Many organizations migrate on-premise software applications to the cloud. However, current coarse-grained cloud migration solutions have made such migrations a non transparent task, an endeavor based on trial-anderror. This paper presents Variability-based, Pattern-driven Architecture Migration .V-PAM), a migration method based on (i) a catalogue of fine-grained service-based cloud architecture migration patterns that target multi-cloud, (ii) a situational migration process framework to guide pattern selection and composition, and (iii) a variability model to structure system migration into a coherent framework. The proposed migration patterns are based on empirical evidence from several migration projects, best practice for cloud architectures and a systematic literature review of existing research. Variability-based, Pattern-driven Architecture Migration allows an organization to (i) select appropriate migration patterns, (ii) compose them to define a migration plan, and (iii) extend them based on the identification of new patterns in new contexts. The patterns are at the core of our solution, embedded into a process model, with their selection governed by a variability model

    Service Isolation vs. Consolidation: Implications for Iaas Cloud Application Deployment

    Get PDF
    Service isolation, achieved by deploying components of multi -tier applications using separate virtual machines (VMs), is a common \u27best\u27 practice. Various advantages cited include simpler deployment architectures, easier resource scalability for supporting dynamic application throughput requirements, and support for component-level fault tolerance . This paper presents results from an empirical study which investigates the performance implications of component placement for deployments of multi -tier applications to Infrastructure-as-a- Service (IaaS) clouds. Relationship s between performance and resource utilization (CPU, disk, network) are investigated to better understand the implications which result from how applications are deployed. All possible deployments for two variants of a multi -tier application were tested, one computationally bound by the model, the other bound by a geospatial database. The best performing deployments required as few as 2 VMs, half the number required for service isolation, demonstrating potential cost savings with service consolidation. Resource use (CPU time, disk I/O, and network I/O) varied based on component placement and VM memory allocation. Using separate VMs to host each application component resulted in performance overhead of ~1 -2%. Relationships between resource utilization an d performance were harnessed to build a multiple linear regression model to predict performance of component deployments. CPU time, disk sector reads, and disk sector writes are identified as the most powerful performance predictors for component deployments

    Migrating on -premises application to windows azure platform (microsoft cloud)

    Get PDF
    Legacy systems are usually attached with outdated technologies which over time become a bottleneck for organizations to manage and maintain. Old and poorly utilized architecture make systems run slow and far from expected, however sometimes organization cannot live without those. Renewing application architecture can be considered as an option but it is time consuming and very costly. Cloud computing as an ultimate solution can be proposed to migrate on-premises application to a utilized environment in terms of infrastructure, computing power and virtualization. In addition, it provides a highly available and elastic computing environment which makes organizations to only pay for what they use. In this research, after a brief introduction to main concepts of cloud computing particularly Windows Azure platform (Microsoft Cloud), it is tried to analyze and assess OnePortfolio system developed by Riskk Sdn Bhd, to see whether it is feasible to be moved to Windows Azure. OnePortfolio operates on a SOA architecture comprised of three main components: services, client application and database. Throughout this research, Windows Azure migration lifecycle in compliance with ISO/IEC 14764 international standard is used as the methodology to perform the migration. Once the application is analyzed and migrated to cloud, it is compared to on-premises environment to evaluate its performance and security mechanism

    Sociomateriality Implications of Software As a Service Adoption on IT-workers’ Roles and Changes in Organizational Routines of IT Systems Support

    Get PDF
    This paper aims to deepen our understanding on how sociomateriality practices influence IT workers’ roles and skill set requirements and changes to the organizational routines of IT systems support, when an organization migrates an on-premise IT system to a software as a service (SaaS) model. This conceptual paper is part of an ongoing study investigating organizations that migrated on-premise IT email systems to SaaS business models, such as Google Apps for Education (GAE) and Microsoft Office 365 systems, in New Zealand tertiary institutions. We present initial findings from interpretive case studies. The findings are, firstly, technological artifacts are entangled in sociomaterial practices, which change the way humans respond to the performative aspects of the organizational routines. Human and material agencies are interwoven in ways that reinforce or change existing routines. Secondly, materiality, virtual realm and spirit of the technology provide elementary levels at which human and material agencies entangle. Lastly, the elementary levels at which human and material entangle depends on the capabilities or skills set of an individual
    corecore