1,928 research outputs found

    Quantifying cloud performance and dependability:Taxonomy, metric design, and emerging challenges

    Get PDF
    In only a decade, cloud computing has emerged from a pursuit for a service-driven information and communication technology (ICT), becoming a significant fraction of the ICT market. Responding to the growth of the market, many alternative cloud services and their underlying systems are currently vying for the attention of cloud users and providers. To make informed choices between competing cloud service providers, permit the cost-benefit analysis of cloud-based systems, and enable system DevOps to evaluate and tune the performance of these complex ecosystems, appropriate performance metrics, benchmarks, tools, and methodologies are necessary. This requires re-examining old system properties and considering new system properties, possibly leading to the re-design of classic benchmarking metrics such as expressing performance as throughput and latency (response time). In this work, we address these requirements by focusing on four system properties: (i) elasticity of the cloud service, to accommodate large variations in the amount of service requested, (ii) performance isolation between the tenants of shared cloud systems and resulting performance variability, (iii) availability of cloud services and systems, and (iv) the operational risk of running a production system in a cloud environment. Focusing on key metrics for each of these properties, we review the state-of-the-art, then select or propose new metrics together with measurement approaches. We see the presented metrics as a foundation toward upcoming, future industry-standard cloud benchmarks

    Trustworthy autonomic architecture (TAArch): Implementation and empirical investigation

    Get PDF
    This paper presents a new architecture for trustworthy autonomic systems. This trustworthy autonomic architecture is different from the traditional autonomic computing architecture and includes mechanisms and instrumentation to explicitly support run-time self-validation and trustworthiness. The state of practice does not lend itself robustly enough to support trustworthiness and system dependability. For example, despite validating system's decisions within a logical boundary set for the system, there’s the possibility of overall erratic behaviour or inconsistency in the system emerging for example, at a different logical level or on a different time scale. So a more thorough and holistic approach, with a higher level of check, is required to convincingly address the dependability and trustworthy concerns. Validation alone does not always guarantee trustworthiness as each individual decision could be correct (validated) but overall system may not be consistent and thus not dependable. A robust approach requires that validation and trustworthiness are designed in and integral at the architectural level, and not treated as add-ons as they cannot be reliably retro-fitted to systems. This paper analyses the current state of practice in autonomic architecture, presents a different architectural approach for trustworthy autonomic systems, and uses a datacentre scenario as the basis for empirical analysis of behaviour and performance. Results show that the proposed trustworthy autonomic architecture has significant performance improvement over existing architectures and can be relied upon to operate (or manage) almost all level of datacentre scale and complexity

    Dynamic bandwidth allocation in multi-class IP networks using utility functions.

    Get PDF
    PhDAbstact not availableFujitsu Telecommunications Europe Lt

    Resource allocation for multimedia messaging services over EGPRS

    Get PDF
    The General Packet Radio Service (GPRS) is a new bearer service for GSM that greatly simplifies wireless access to packet data networks, such as the Internet, corporate LANs or to mobile portals. It applies a packet radio standard to transfer user data packets in wellorganized way between Mobile Stations (MS) and external packet data networks. The Enhanced General Packet Radio Service (EGPRS) is an extension of GPRS, offering much greater capacity. These enhancements have allowed the introduction of new services like Multimedia Messaging Services (MMS). MMS enables messaging with full content versatility, including images, audio, video, data and text, from terminal to terminal or from terminal to e-mail. The Wireless Application Protocol (WAP) is the WAP Forum standard for the presentation and delivery of wireless information and telephony services on mobile phones and other wireless terminals. In this thesis it is indicated that efficient radio resource allocation is necessary for managing different types of traffic in order to maintain the quality demands for different types of services. A theoretical model of MMS and WAP traffic is developed, and based on this model a simulator is implemented in Java programming language. This thesis proposes two techniques to improve the radio resource allocation algorithm performance called "radio link condition diversification" and "interactive traffic class prioritization". The radio link condition diversification technique defines minimum radio link quality that allows the user to receive their packets. The interactive traffic class prioritization technique defines different priorities for WAP packets and for MMS packets. Both techniques give good results in increasing user's perception of services and increasing network efficiency. This thesis indicates also that the prioritization mechanism successfully improves the response time of the interactive service by up to 80% with a setting of priority for interactive traffic class and decreasing the performance of the background traffic. This decrease is within a range acceptable by the end-user and that the link conditions limit mechanism has an advantage in terms of resource utilization

    Adaptive Quality of Service Control in Distributed Real-Time Embedded Systems

    Get PDF
    An increasing number of distributed real-time embedded systems face the critical challenge of providing Quality of Service (QoS) guarantees in open and unpredictable environments. For example, such systems often need to enforce CPU utilization bounds on multiple processors in order to avoid overload and meet end-to-end dead-lines, even when task execution times deviate signiïŹcantly from their estimated values or change dynamically at run-time. This dissertation presents an adaptive QoS control framework which includes a set of control design methodologies to provide robust QoS assurance for systems at diïŹ€erent scales. To demonstrate its eïŹ€ectiveness, we have applied the framework to the end-to-end CPU utilization control problem for a common class of distributed real-time embedded systems with end-to-end tasks. We formulate the utilization control problem as a constrained multi-input-multi-output control model. We then present a centralized control algorithm for small or medium size systems, and a decentralized control algorithm for large-scale systems. Both algorithms are designed systematically based on model predictive control theory to dynamically enforce desired utilizations. We also introduce novel task allocation algorithms to ensure that the system is controllable and feasible for utilization control. Furthermore, we integrate our control algorithms with fault-tolerance mechanisms as an eïŹ€ective way to develop robust middleware systems, which maintain both system reliability and real-time performance even when the system is in face of malicious external resource contentions and permanent processor failures. Both control analysis and extensive experiments demonstrate that our control algorithms and middleware systems can achieve robust utilization guarantees. The control framework has also been successfully applied to other distributed real-time applications such as end-to-end delay control in real-time image transmission. Our results show that adaptive QoS control middleware is a step towards self-managing, self-healing and self-tuning distributed computing platform

    QoE on media deliveriy in 5G environments

    Get PDF
    231 p.5G expandirĂĄ las redes mĂłviles con un mayor ancho de banda, menor latencia y la capacidad de proveer conectividad de forma masiva y sin fallos. Los usuarios de servicios multimedia esperan una experiencia de reproducciĂłn multimedia fluida que se adapte de forma dinĂĄmica a los intereses del usuario y a su contexto de movilidad. Sin embargo, la red, adoptando una posiciĂłn neutral, no ayuda a fortalecer los parĂĄmetros que inciden en la calidad de experiencia. En consecuencia, las soluciones diseñadas para realizar un envĂ­o de trĂĄfico multimedia de forma dinĂĄmica y eficiente cobran un especial interĂ©s. Para mejorar la calidad de la experiencia de servicios multimedia en entornos 5G la investigaciĂłn llevada a cabo en esta tesis ha diseñado un sistema mĂșltiple, basado en cuatro contribuciones.El primer mecanismo, SaW, crea una granja elĂĄstica de recursos de computaciĂłn que ejecutan tareas de anĂĄlisis multimedia. Los resultados confirman la competitividad de este enfoque respecto a granjas de servidores. El segundo mecanismo, LAMB-DASH, elige la calidad en el reproductor multimedia con un diseño que requiere una baja complejidad de procesamiento. Las pruebas concluyen su habilidad para mejorar la estabilidad, consistencia y uniformidad de la calidad de experiencia entre los clientes que comparten una celda de red. El tercer mecanismo, MEC4FAIR, explota las capacidades 5G de analizar mĂ©tricas del envĂ­o de los diferentes flujos. Los resultados muestran cĂłmo habilita al servicio a coordinar a los diferentes clientes en la celda para mejorar la calidad del servicio. El cuarto mecanismo, CogNet, sirve para provisionar recursos de red y configurar una topologĂ­a capaz de conmutar una demanda estimada y garantizar unas cotas de calidad del servicio. En este caso, los resultados arrojan una mayor precisiĂłn cuando la demanda de un servicio es mayor

    Cooperative resource management in the cloud

    Get PDF
    L’évolution des infrastructures informatiques encourage la gestion sĂ©parĂ©e de l’infrastructure matĂ©rielle et de celle des logiciels. Dans cette direction, les infrastructures de cloud virtualisĂ©es sont devenues trĂ©s populaires. Parmi les diffĂ©rents modĂšles de cloud, les Infrastructures as a Service (IaaS) ont de nombreux avantages pour le fournisseur comme pour le client. Dans ce modĂšle de cloud, le fournisseur fournit ses ressources virtualisĂ©es et il est responsable de la gestion de son infrastructure. De son cotĂ©, le client gĂšre son application qui est dĂ©ployĂ©e dans les machines virtuelles allouĂ©es. Ces deux acteurs s’appuient gĂ©nĂ©ralement sur des systĂšmes d’administration autonomes pour automatiser les tĂąches d’administration. RĂ©duire la quantitĂ© de ressources utilisĂ©es (et la consommation d’énergie) est un des principaux objectifs de ce modĂšle de cloud. Cette rĂ©duction peut ĂȘtre obtenue Ă  l’exĂ©cution au niveau de l’application par le client (en redimensionnant l’application) ou au niveau du systĂšme virtualisĂ© par le fournisseur (en regroupant les machines virtuelles dans l’infrastructure matĂ©rielle en fonction de leur charge). Dans les infrastructures de cloud traditionnelles, les politiques de gestion de ressources ne sont pas coopĂ©ratives : le fournisseur ne possĂšde pas d’informations dĂ©taillĂ©es sur les applications. Ce manque de coordination engendre des surcoĂ»ts et des gaspillages de ressources qui peuvent ĂȘtre rĂ©duits avec une politique de gestion de ressources coopĂ©rative. Dans cette thĂšse, nous traitons du problĂšme de la gestion de ressources sĂ©parĂ©e dans un environnement de cloud virtualisĂ©. Nous proposons un modĂšle de machines virtuelles Ă©lastiques avec une politique de gestion coopĂ©rative des ressources. Cette politique associe la connaissance des deux acteurs du cloud afin de rĂ©duire les coĂ»ts et la consommation d’énergie. Nous Ă©valuons les bĂ©nĂ©fices de cette approche avec plusieurs expĂ©riences dans un IaaS privĂ©. Cette Ă©valuation montre que notre politique est meilleure que la gestion des ressources non coordonnĂ©e dans un IaaS traditionnel, car son impact sur les performances est faible et elle permet une meilleure utilisation des ressources matĂ©rielles et logicielles. ABSTRACT : Recent advances in computer infrastructures encourage the separation of hardware and software management tasks. Following this direction, virtualized cloud infrastructures are becoming very popular. Among various cloud models, Infrastructure as a Service (IaaS) provides many advantages to both provider and customer. In this service model, the provider offers his virtualized resource, and is responsible for managing his infrastructure, while the customer manages his application deployed in the allocated virtual machines. These two actors typically use autonomic resource management systems to automate these tasks at runtime. Minimizing the amount of resource (and power consumption) in use is one of the main services that such cloud model must ensure. This objective can be done at runtime either by the customer at the application level (by scaling the application) or by the provider at the virtualization level (by migrating virtual machines based on the infrastructure’s utilization rate). In traditional cloud infrastructures, these resource management policies work uncoordinated: knowledge about the application is not shared with the provider. This behavior faces application performance overheads and resource wasting, which can be reduced with a cooperative resource management policy. In this research work, we discuss the problem of separate resource management in the cloud. After having this analysis, we propose a direction to use elastic virtual machines with cooperative resource management. This policy combines the knowledge of the application and the infrastructure in order to reduce application performance overhead and power consumption. We evaluate the benefit of our cooperative resource management policy with a set of experiments in a private IaaS. The evaluation shows that our policy outperforms uncoordinated resource management in traditional IaaS with lower performance overhead, better virtualized and physical resource usage

    Towards a novel biologically-inspired cloud elasticity framework

    Get PDF
    With the widespread use of the Internet, the popularity of web applications has significantly increased. Such applications are subject to unpredictable workload conditions that vary from time to time. For example, an e-commerce website may face higher workloads than normal during festivals or promotional schemes. Such applications are critical and performance related issues, or service disruption can result in financial losses. Cloud computing with its attractive feature of dynamic resource provisioning (elasticity) is a perfect match to host such applications. The rapid growth in the usage of cloud computing model, as well as the rise in complexity of the web applications poses new challenges regarding the effective monitoring and management of the underlying cloud computational resources. This thesis investigates the state-of-the-art elastic methods including the models and techniques for the dynamic management and provisioning of cloud resources from a service provider perspective. An elastic controller is responsible to determine the optimal number of cloud resources, required at a particular time to achieve the desired performance demands. Researchers and practitioners have proposed many elastic controllers using versatile techniques ranging from simple if-then-else based rules to sophisticated optimisation, control theory and machine learning based methods. However, despite an extensive range of existing elasticity research, the aim of implementing an efficient scaling technique that satisfies the actual demands is still a challenge to achieve. There exist many issues that have not received much attention from a holistic point of view. Some of these issues include: 1) the lack of adaptability and static scaling behaviour whilst considering completely fixed approaches; 2) the burden of additional computational overhead, the inability to cope with the sudden changes in the workload behaviour and the preference of adaptability over reliability at runtime whilst considering the fully dynamic approaches; and 3) the lack of considering uncertainty aspects while designing auto-scaling solutions. This thesis seeks solutions to address these issues altogether using an integrated approach. Moreover, this thesis aims at the provision of qualitative elasticity rules. This thesis proposes a novel biologically-inspired switched feedback control methodology to address the horizontal elasticity problem. The switched methodology utilises multiple controllers simultaneously, whereas the selection of a suitable controller is realised using an intelligent switching mechanism. Each controller itself depicts a different elasticity policy that can be designed using the principles of fixed gain feedback controller approach. The switching mechanism is implemented using a fuzzy system that determines a suitable controller/- policy at runtime based on the current behaviour of the system. Furthermore, to improve the possibility of bumpless transitions and to avoid the oscillatory behaviour, which is a problem commonly associated with switching based control methodologies, this thesis proposes an alternative soft switching approach. This soft switching approach incorporates a biologically-inspired Basal Ganglia based computational model of action selection. In addition, this thesis formulates the problem of designing the membership functions of the switching mechanism as a multi-objective optimisation problem. The key purpose behind this formulation is to obtain the near optimal (or to fine tune) parameter settings for the membership functions of the fuzzy control system in the absence of domain experts’ knowledge. This problem is addressed by using two different techniques including the commonly used Genetic Algorithm and an alternative less known economic approach called the Taguchi method. Lastly, we identify seven different kinds of real workload patterns, each of which reflects a different set of applications. Six real and one synthetic HTTP traces, one for each pattern, are further identified and utilised to evaluate the performance of the proposed methods against the state-of-the-art approaches
    • 

    corecore