669 research outputs found

    Strategic and operational services for workload management in the cloud

    Full text link
    In hosting environments such as Infrastructure as a Service (IaaS) clouds, desirable application performance is typically guaranteed through the use of Service Level Agreements (SLAs), which specify minimal fractions of resource capacities that must be allocated by a service provider for unencumbered use by customers to ensure proper operation of their workloads. Most IaaS offerings are presented to customers as fixed-size and fixed-price SLAs, that do not match well the needs of specific applications. Furthermore, arbitrary colocation of applications with different SLAs may result in inefficient utilization of hosts' resources, resulting in economically undesirable customer behavior. In this thesis, we propose the design and architecture of a Colocation as a Service (CaaS) framework: a set of strategic and operational services that allow the efficient colocation of customer workloads. CaaS strategic services provide customers the means to specify their application workload using an SLA language that provides them the opportunity and incentive to take advantage of any tolerances they may have regarding the scheduling of their workloads. CaaS operational services provide the information necessary for, and carry out the reconfigurations mandated by strategic services. We recognize that it could be the case that there are multiple, yet functionally equivalent ways to express an SLA. Thus, towards that end, we present a service that allows the provably-safe transformation of SLAs from one form to another for the purpose of achieving more efficient colocation. Our CaaS framework could be incorporated into an IaaS offering by providers or it could be implemented as a value added proposition by IaaS resellers. To establish the practicality of such offerings, we present a prototype implementation of our proposed CaaS framework

    Dynamic Window-Constrained Scheduling for Real-Time Media Streaming

    Full text link
    This paper describes an algorithm for scheduling packets in real-time multimedia data streams. Common to these classes of data streams are service constraints in terms of bandwidth and delay. However, it is typical for real-time multimedia streams to tolerate bounded delay variations and, in some cases, finite losses of packets. We have therefore developed a scheduling algorithm that assumes streams have window-constraints on groups of consecutive packet deadlines. A window-constraint defines the number of packet deadlines that can be missed in a window of deadlines for consecutive packets in a stream. Our algorithm, called Dynamic Window-Constrained Scheduling (DWCS), attempts to guarantee no more than x out of a window of y deadlines are missed for consecutive packets in real-time and multimedia streams. Using DWCS, the delay of service to real-time streams is bounded even when the scheduler is overloaded. Moreover, DWCS is capable of ensuring independent delay bounds on streams, while at the same time guaranteeing minimum bandwidth utilizations over tunable and finite windows of time. We show the conditions under which the total demand for link bandwidth by a set of real-time (i.e., window-constrained) streams can exceed 100% and still ensure all window-constraints are met. In fact, we show how it is possible to guarantee worst-case per-stream bandwidth and delay constraints while utilizing all available link capacity. Finally, we show how best-effort packets can be serviced with fast response time, in the presence of window-constrained traffic

    Renegotiation based dynamic bandwidth allocation for selfsimilar VBR traffic

    Get PDF
    The provision of QoS to applications traffic depends heavily on how different traffic types are categorized and classified, and how the prioritization of these applications are managed. Bandwidth is the most scarce network resource. Therefore, there is a need for a method or system that distributes an available bandwidth in a network among different applications in such a way that each class or type of traffic receives their constraint QoS requirements. In this dissertation, a new renegotiation based dynamic resource allocation method for variable bit rate (VBR) traffic is presented. First, pros and cons of available off-line methods that are used to estimate selfsimilarity level (represented by Hurst parameter) of a VBR traffic trace are empirically investigated, and criteria to select measurement parameters for online resource management are developed. It is shown that wavelet analysis based methods are the strongest tools in estimation of Hurst parameter with their low computational complexities, compared to the variance-time method and R/S pox plot. Therefore, a temporal energy distribution of a traffic data arrival counting process among different frequency sub-bands is considered as a traffic descriptor, and then a robust traffic rate predictor is developed by using the Haar wavelet analysis. The empirical results show that the new on-line dynamic bandwidth allocation scheme for VBR traffic is superior to traditional dynamic bandwidth allocation methods that are based on adaptive algorithms such as Least Mean Square, Recursive Least Square, and Mean Square Error etc. in terms of high utilization and low queuing delay. Also a method is developed to minimize the number of bandwidth renegotiations to decrease signaling costs on traffic schedulers (e.g. WFQ) and networks (e.g. ATM). It is also quantified that the introduced renegotiation based bandwidth management scheme decreases heavytailedness of queue size distributions, which is an inherent impact of traffic self similarity. The new design increases the achieved utilization levels in the literature, provisions given queue size constraints and minimizes the number of renegotiations simultaneously. This renegotiation -based design is online and practically embeddable into QoS management blocks, edge routers and Digital Subscriber Lines Access Multiplexers (DSLAM) and rate adaptive DSL modems

    Taming Energy Costs of Large Enterprise Systems Through Adaptive Provisioning

    Get PDF
    One of the most pressing concerns in modern datacenter management is the rising cost of operation. Therefore, reducing variable expense, such as energy cost, has become a number one priority. However, reducing energy cost in large distributed enterprise system is an open research topic. These systems are commonly subjected to highly volatile workload processes and characterized by complex performance dependencies. This paper explicitly addresses this challenge and presents a novel approach to Taming Energy Costs of Larger Enterprise Systems (Tecless). Our adaptive provisioning methodology combines a low-level technical perspective on distributed systems with a high-level treatment of workload processes. More concretely, Tecless fuses an empirical bottleneck detection model with a statistical workload prediction model. Our methodology forecasts the system load online, which enables on-demand infrastructure adaption while continuously guaranteeing quality of service. In our analysis we show that the prediction of future workload allows adaptive provisioning with a power saving potential of up 25 percent of the total energy cost

    Investigation of delay jitter of heterogeneous traffic in broadband networks

    Get PDF
    Scope and Methodology of Study: A critical challenge for both wired and wireless networking vendors and carrier companies is to be able to accurately estimate the quality of service (QoS) that will be provided based on the network architecture, router/switch topology, and protocol applied. As a result, this thesis focuses on the theoretical analysis of QoS parameters in term of inter-arrival jitter in differentiated services networks by deploying analytic/mathematical modeling technique and queueing theory, where the analytic model is expressed in terms of a set of equations that can be solved to yield the desired delay jitter parameter. In wireless networks with homogeneous traffic, the effects on the delay jitter in reference to the priority control scheme of the ARQ traffic for the two cases of: 1) the ARQ traffic has a priority over the original transmission traffic; and 2) the ARQ traffic has no priority over the original transmission traffic are evaluated. In wired broadband networks with heterogeneous traffic, the jitter analysis is conducted and the algorithm to control its effect is also developed.Findings and Conclusions: First, the results show that high priority packets always maintain the minimum inter-arrival jitter, which will not be affected even in heavy load situation. Second, the Gaussian traffic modeling is applied using the MVA approach to conduct the queue length analysis, and then the jitter analysis in heterogeneous broadband networks is investigated. While for wireless networks with homogeneous traffic, binomial distribution is used to conduct the queue length analysis, which is sufficient and relatively easy compared to heterogeneous traffic. Third, develop a service discipline called the tagged stream adaptive distortion-reducing peak output-rate enforcing to control and avoid the delay jitter increases without bound in heterogeneous broadband networks. Finally, through the analysis provided, the differential services, was proved not only viable, but also effective to control delay jitter. The analytic models that serve as guidelines to assist network system designers in controlling the QoS requested by customer in term of delay jitter

    Fixed-Priority Scheduling Algorithms with Multiple Objectives in Hard Real-Time Systems

    Get PDF
    In the context ofFixed-Priority Scheduling in Real-Time Systems, we investigate scheduling mechanisms for supporting systems where, in addition to timing constraints, their performance with respect to additional QoS requirements must be improved. This'type of situation may occur when the worst-case res~urce requirements of all or some running tasks cannot be simultaneously met due to task contention. . Solutions to these problems have been proposed in the context of both fixed-priority and dynamic-priority scheduling. In fixed-priority scheduling, the typical approach is to artificially modify the attributes or structure of tasks, and/or usually require non-standard run-time support. In dynamic-priority scheduling approaches, utility functions are employed to make scheduling decisions with the objective of maximising the utility. The main difficulties with these approaches are the inability to formulate and model appropriately utility functions for each task, and the inability to guarantee hard deadlines without executing computationally costly algorithms. In this thesis we propose a different approach. Firstly, we introduce the concept of relative importance among tasks as a new metric for expressing QoS requirements. The meaning of this importance relationship is to express that in a schedule it i~ desirable to run a task in preference to other ones. This model is more intuitive and less restrictive than traditional utility-based app~oaches. Secondly, we formulate a scheduling problem in terms of finding a feasible assignment of fixed priorities that maximises the new QoS metric, and propose the DI and DI+ algorithms that find optimal solutions. By extensive simulation, we show that the new QoS metric combined with the DI algorithm outperforms the rate monotonic priority algorithm in several practical problems such as minimising jitter, minimising the number of preemptions or minimising the latency. In addition, our approach outperforms EDF in several scenarios

    Analysis and evaluation of in-home networks based on HomePlug-AV power line communications

    Get PDF
    [ESP] No hace mucho tiempo, las redes in-home (también denominadas redes domésticas) únicamente se utilizaban para interconectar los diferentes ordenadores de una vivienda, de manera que pudieran compartir una impresora entre ellos. Hoy en día, sin embargo, esta definición es mucho más amplia debido a la gran cantidad de dispositivos existentes en la vivienda con capacidad de conectarse a una red para transmitir y recibir información. En una red in-home actual, podemos encontrar desde teléfonos móviles equipados con conectividad WI-FI a dispositivos NAS (Network Attached Storage), utilizados para almacenar información, imágenes o videos en red, que a su vez pueden ser transferidos a televisiones de alta definición u ordenadores. A la hora de instalar una red de comunicaciones en una vivienda, se persiguen principalmente dos objetivos, reducir el coste de instalación y conseguir una gran flexibilidad de cara a futuras ampliaciones. Una red basada en tecnología PLC (Power Line Communications) cumple estos requisitos ya que, al utilizar la infraestructura de cableado eléctrico existente en la vivienda, es muy sencilla y económica de instalar y ampliar. Dentro de la tecnología PLC existen diferentes estándares, siendo HomePlug-AV (HomePlug Audio-Video o simplemente HPAV) el más extendido en la actualidad para la instalación de redes domésticas. Este estándar permite alcanzar velocidades de transmisión de hasta 200Mbps a través de los cables de baja tensión de una vivienda convencional. El objetivo principal de esta tesis doctoral es aportar nuevas ideas que mejoren las prestaciones de las redes in-home basadas en la tecnología PLC, utilizando como base el estándar Homeplug-AV. Estas redes utilizan una arquitectura centralizada, en la que la mayor parte de la inteligencia de red está concentrada en un coordinador central (CCo, por sus siglas en inglés). Por lo tanto, la mayor parte de las modificaciones propuestas irán encaminadas a mejorar dicho dispositivo, que podrá llegar a convertirse en un gestor de red capaz de manejar conjuntamente interfaces de diferentes tecnologías. En primer lugar, se presenta un análisis detallado del comportamiento del estándar en diferentes situaciones que se pueden producir de manera común en una red doméstica. Este análisis se realizó tanto con dispositivos reales como mediante simulación. Para el segundo tipo de medidas, se diseñó un simulador de la tecnología HomePlug que implementa el nivel físico y el nivel MAC de la misma, junto con modelos de los servicios más utilizados en entornos domésticos. Este simulador se utilizó tanto para estas medidas iniciales como para evaluar las diferentes modificaciones del estándar propuestas posteriormente en este trabajo. Este análisis proporcionó dos resultados significativos. En primer lugar, se comprobó que al introducir un modelo real de nivel físico al protocolo CSMA/CA utilizado a nivel MAC se producían resultados muy diferentes a los presentados en los modelos publicados hasta ese momento. Por ello, se propuso un modelo matemático que incorporaba dichos efectos. En segundo lugar, se identificaron diferentes áreas de la tecnología que eran susceptibles de mejora. El resto de la tesis se centró entonces en la mejora de dichos puntos débiles. El primero de estos puntos débiles está relacionado con las transmisión de datos unicast. El medio PLC es selectivo en frecuencia y muy dependiente del tiempo y de la localización de las estaciones. Incluso es posible que, en un mismo enlace, la capacidad de los enlaces ascendente y descendente sea distinta. En estos entornos, la utilización del protocolo de transporte TCP presenta serios problemas, ya que define gran parte de sus parámetros en función del Round Trip time (RTT) del enlace. Como alternativa se pensó en los códigos Fountain. Este tipo de codificación de fuente permite realizar transmisiones fiables de datos sin necesidad de utilizar un canal de retorno, evitando de esta forma los problemas derivados de las asimetrías de la red. Se realizaron varios experimentos comparando ambas soluciones, y se comprobó que las prestaciones de este tipo de codificaciones superan al protocolo TCP a la hora de transmitir ficheros de manera fiable a través de las redes PLC. Además, los códigos Fountain también se utilizaron para el diseño de otra aplicación. Es muy común que en un escenario doméstico haya disponible más de una tecnología (Wi-Fi, Ethernet, PLC, etc). Tenemos por tanto que una aplicación capaz de integrar interfaces de diferentes tecnologías podría ser muy útil en estos entornos, ya que se podría conseguir un mayor ancho de banda, mayor tolerancia a errores, balanceo de carga, etc. El kernel de Linux dispone de un módulo denominado Bonding que permite agrupar diferentes interfaces Ethernet. Sin embargo, no está preparado para agrupar interfaces de diferentes tecnologías, y mucho menos para tecnologás de capacidad variable como es el caso de PLC o de las comunicaciones inalámbricas. Por ello, se realizó una modificación de dicho driver utilizando para ello los códigos Fountain, que solucionan los problemas que se pueden producir debido a las variaciones de capacidad. Por otra parte, con la actual versión del estándar HomePlug AV, las comunicaciones multicast presentan unas prestaciones muy pobres. Esto es debido a que, a pesar de que el canal PLC es broadcast, la naturaleza de la modulación OFDM (Ortogonal Frequency Division Multiplexing) que se utiliza a nivel físico es punto a punto. Esto hace que las transmisiones simultáneas a un grupo de receptores se traduzcan automáticamente en sucesivas transmisiones punto a punto a los diferentes miembros del grupo. Con esta técnica, la capacidad efectiva de transmisión multicast disminuye de manera muy importante a medida que aumenta el número de receptores. En este trabajo se han propuesto dos técnicas alternativas. La primera consiste en la utilización de un mapa de tonos común para todos los miembros del grupo multicast, asignado a estas comunicaciones los parámetros de modulación del cliente con las peores condiciones de canal. Este algoritmo ha sido tradicionalmente descartado en los sistemas OFDM por sus bajas prestaciones. Sin embargo, la correlación existente entre los diferentes canales de una red PLC hace que su comportamiento sea mucho mejor. Además, se propuso un segundo algoritmo que utilizaba técnicas de optimización para maximizar la tasa de comunicación multicast, obteniendo un mejor comportamiento cuando el número de clientes es elevado. Por último, en redes de capacidad física variable, como es el caso de las redes PLC, las técnicas cross-layer están despertando un gran interés. Este tipo de algoritmos están basado en la compartición de información entre diferentes capas de la estructura OSI para mejorar el comportamiento del sistema. En este trabajo se ha propuesto un algoritmo que modifica los parámetros del protocolo CSMA/CA de nivel MAC utilizando información de nivel físico y los requerimientos de QoS del servicio de niveles superiores. De esta forma se consigue dar prioridad en el acceso al medio a los clientes con problemas de QoS, mejorando de esta forma del comportamiento de la red. Este algoritmo ha sido evaluado mediante simulación en un escenario doméstico típico, comprobando que ofrece unos resultados muy prometedores. [ENG] Not very long time ago, in-home networks (also called domestic networks) were only used to share a printer between a number of computers. Nowadays, however, due to the huge amount of devices present at home with communication capabilities, this definition has become much wider. In a current in-home network we can find, from mobile phones with wireless connectivity, or NAS (Network Attached Storage) devices sharing multimedia content with high-definition televisions or computers. When installing a communications network in a home, two objectives are mainly pursued: Reducing cost and high flexibility in supporting future network requirements. A network based on Power Line Communications (PLC) technology is able to fulfill these objectives, since as it uses the low voltage wiring already available at home, it is very easy to install and expand, providing a cost-effective solution for home environments. There are different PLC standards, being HomePlug-AV (HomePlug Audio-Video, or simply HPAV) the most widely used nowadays. This standard is able to achieve transmission rates up to 200 Mpbs through the electrical wiring of a typical home. The main objective of this thesis is to provide new ideas to improve the performance of PLC technology based in-home networks, using as starting point the HPAV standard. A network based on this technology uses a centralized architecture, in which the most important part of the network intelligence is concentrated in a single device, the Central Coordinator (CCo). Hence, most of the modifications proposed in this work will try to improve this particular device, which can even become a multi-technology central manager, able to combine interfaces of different technologies to improve the network performance. Initially, it is presented a detailed analysis of HPAV performance in some scenarios typically found in a home environment. It was done through simulation and by experimentation using real devices. To obtain the former results, it was designed a HPAV simulator which implements the physical (PHY) and medium access control (MAC) layers of the standard, together with a traffic modeling module which implements the services most commonly found in a home network. This simulation tool was used both in these initial measurements and to evaluate the standard modifications that are proposed in this work. This analysis provides two main results. Firstly, it was found that when a real PHY model is used together with the CSMA/CA MAC protocol the simulation results were very different to those obtained with previously presented mathematical models of this protocol. Hence, it was proposed a new model that considers these effects. Next, some areas of the technology which could be improved were identified. The rest of the thesis was then centered around proposing solutions to these weaknesses. The first weakness solved is related to unicast data transmission. PLC medium is frequency selective and time variant, and it presents a remarkable variation among locations or depending on the connected loads. Even in a single link, the channel capacities between transmitter and receiver can be very asymmetric. In such environments, the use of TCP as transport protocol presents serious problems, since it defines some of its parameters according to the Round Trip Time (RTT). Alternatively, the use of Fountain codes for reliable data transmission in these environments was proposed. These codes allow to transmit information without a feedback channel, overcoming in this way the problems related to the variability of the channel. Different experiments were performed comparing both solutions, concluding that in PLC based networks the performance achieved by Fountain codes outperforms the results obtained with a TCP-based application. In addition, Fountain codes were also used for another application. In home environments, it is very common to find more than one available technology to deploy a network (Wi-Fi, Ethernet, PLC, etc). Therefore, an application that makes possible the aggregation of different interfaces would be very useful, as it will provide higher bandwidth, fault tolerance and load balancing. The Linux Kernel contains a driver (Bonding) which allows Ethernet interfaces aggregation. However, it is not prepared for asymmetric interfaces aggregation and even less for variable capacity technologies like PLC or Wi-Fi. In this work, it is presented a modification of this driver which uses Fountain codes to solve the problems that may arise when asymmetric interfaces are aggregated. On another note, multicast communications in the actual HPAV standard versions presents serious problems. This is because, although PLC medium is broadcast by nature, the Orthogonal Frequency Division Multiplexing (OFDM) modulation used at PHY layer is always point to point. Therefore, multicast communications are carried out as successive point-to-point transmissions to the different members of the group. This technique clearly degrades the performance of multicast services as the number of receivers increases. In this work, they have been proposed two alternative algorithms. The first one consists of using a common tone map for all the multicast group members. This tone map corresponds to the modulation parameters obtained for the client with the worst channel conditions. This algorithm has been traditionally discarded in OFDM systems because of its poor performance. However, in contrast to other technologies (like wireless for example), channel responses in a given PLC network exhibit significant correlation among them. This reduces the differences among the users, improving the performance of this algorithm. In addition, another technique which uses an optimization algorithm to maximize the multicast bit rate is also evaluated, obtaining that its use can be suitable when the number of multicast clients is high. Finally, due to the properties of PLC medium, cross-layer technique are eliciting a big interest. These algorithms are based in the information sharing between adjacent layers in the OSI model to improve the system behavior. In this work, it has been proposed an extension of the HPAV CSMA/CA algorithm which modifies the protocol parameters using PHY layer information and the QoS requirements of the upper-layer services. In this way, priority access to the channel can be provided to the nodes with QoS problems, improving the whole network performance. This algorithm has been evaluated through simulation in a typical home environment with very promising results.Universidad Politécnica de Cartagen
    corecore