1,007 research outputs found

    Datacenter Traffic Control: Understanding Techniques and Trade-offs

    Get PDF
    Datacenters provide cost-effective and flexible access to scalable compute and storage resources necessary for today's cloud computing needs. A typical datacenter is made up of thousands of servers connected with a large network and usually managed by one operator. To provide quality access to the variety of applications and services hosted on datacenters and maximize performance, it deems necessary to use datacenter networks effectively and efficiently. Datacenter traffic is often a mix of several classes with different priorities and requirements. This includes user-generated interactive traffic, traffic with deadlines, and long-running traffic. To this end, custom transport protocols and traffic management techniques have been developed to improve datacenter network performance. In this tutorial paper, we review the general architecture of datacenter networks, various topologies proposed for them, their traffic properties, general traffic control challenges in datacenters and general traffic control objectives. The purpose of this paper is to bring out the important characteristics of traffic control in datacenters and not to survey all existing solutions (as it is virtually impossible due to massive body of existing research). We hope to provide readers with a wide range of options and factors while considering a variety of traffic control mechanisms. We discuss various characteristics of datacenter traffic control including management schemes, transmission control, traffic shaping, prioritization, load balancing, multipathing, and traffic scheduling. Next, we point to several open challenges as well as new and interesting networking paradigms. At the end of this paper, we briefly review inter-datacenter networks that connect geographically dispersed datacenters which have been receiving increasing attention recently and pose interesting and novel research problems.Comment: Accepted for Publication in IEEE Communications Surveys and Tutorial

    Economic FAQs About the Internet

    Get PDF
    This is a set of Frequently Asked Questions (and answers) about the economic, institutional, and technological structure of the Internet. We describe the history and current state of the Internet, discuss some of the pressing economic and regulatory problems, and speculate about future developments.Internet, telecommunications, congestion pricing, National Information Infrastructure

    On the Load Balancing of Edge Computing Resources for On-Line Video Delivery

    Get PDF
    Online video broadcasting platforms are distributed, complex, cloud oriented, scalable, micro-service-based systems that are intended to provide over-the-top and live content to audience in scattered geographic locations. Due to the nature of cloud VM hosting costs, the subscribers are usually served under limited resources in order to minimize delivery budget. However, operations including transcoding require high-computational capacity and any disturbance in supplying requested demand might result in quality of experience (QoE) deterioration. For any online delivery deployment, understanding user's QoE plays a crucial role for rebalancing cloud resources. In this paper, a methodology for estimating QoE is provided for a scalable cloud-based online video platform. The model will provide an adeptness guideline regarding limited cloud resources and relate computational capacity, memory, transcoding and throughput capability, and finally latency competence of the cloud service to QoE. Scalability and efficiency of the system are optimized through reckoning sufficient number of VMs and containers to satisfy the user requests even on peak demand durations with minimum number of VMs. Both horizontal and vertical scaling strategies (including VM migration) are modeled to cover up availability and reliability of intermediate and edge content delivery network cache nodes

    The Design of Medium Access Control (MAC) Protocols for Energy Efficient and QoS Provision in Wireless Sensor Networks

    Get PDF
    This thesis work focuses on innovative design of media access control (MAC) protocols in wireless sensor networks (WNSs). The characteristics of the WSN inquire that the network service design considers both energy efficiency and the associated application requirement. However, most existing protocols address only the issue of energy efficiency. In this thesis, a MAC protocol has been proposed (referred to as Q-MAC) that not only minimized the energy consumption in multi-hop WSNs, but also provides Quality of Service (QoS) by differentiating network services based on priority levels prescribed by different applications. The priority levels reflect the state of system resources including residual energy and queue occupancies. Q-MAC contains both intra- and inter- node arbitration mechanisms. The intra-node packet scheduling employs a multiple queuing architectures, and applies a scheduling scheme consisting of packet classification and weighted arbitration. We introduce the Power Conservation MACAW (PC-MACAW), a power-aware scheduling mechanism which, together with the Loosely Prioritized Random Access (LPRA) algorithm, govern the inter-node scheduling. Performance evaluation are conducted between Q-MAC and S-MAC with respect to two performance metrics: energy consumption and average latency. Simulation results indicate Q-MAC achieves comparable performance to that of S-MAC in non-prioritized traffic scenarios. When packets with different priorities are introduced, Q-MAC yields noticeable average latency differentiations between the classes of service, while preserving the same degree of energy consumption as that of S-MAC. Since the high density nature of WSN may introduce heavy traffic load and thus consume large amount of energy for communication, another MAC protocol, referred to as the Deployment-oriented MAC (D-MAC)has been further proposed. D-MAC minimalizes both sensing and communication redundancy by putting majority of redundant nodes into the sleep state. The idea is to establish a sensing and communication backbone covering the whole sensing field with the least sensing and communication redundancy. In specific, we use equal-size rectangular cells to partition the sensing field and chose the size of each cell in a way such that regardless of the actual location within the cell, a node can always sense the whole cell and communicate with all the nodes in neighboring cells. Once the sensing field has been partitioned using these cells, a localized Location-aware Selection Algorithm (LSA) is carried out to pick up only one node within each cell to be active for a fixed amount of period. This selection is energy-oriented, only nodes with a maximum energy will be on and the rest of nodes will be put into the sleep state once the selection process is over. To balance the energy consumption, the selection algorithm is periodically conducted until all the nodes are out of power. Simulation results indicated that D-MAC saves around 80% energy compared to that of S-MAC and Q-MAC, while maintaining 99% coverage. D-MAC is also superior to S-MAC and Q-MAC in terms of average latency. However, the use of GPS in D-MAC in identifying the nodes within the same cell, would cause extra cost and complexity for the design of sensor nodes

    Application acceleration for wireless and mobile data networks

    Get PDF
    This work studies application acceleration for wireless and mobile data networks. The problem of accelerating application can be addressed along multiple dimensions. The first dimension is advanced network protocol design, i.e., optimizing underlying network protocols, particulary transport layer protocol and link layer protocol. Despite advanced network protocol design, in this work we observe that certain application behaviors can fundamentally limit the performance achievable when operating over wireless and mobile data networks. The performance difference is caused by the complex application behaviors of these non-FTP applications. Explicitly dealing with application behaviors can improve application performance for new environments. Along this overcoming application behavior dimension, we accelerate applications by studying specific types of applications including Client-server, Peer-to-peer and Location-based applications. In exploring along this dimension, we identify a set of application behaviors that significantly affect application performance. To accommodate these application behaviors, we firstly extract general design principles that can apply to any applications whenever possible. These design principles can also be integrated into new application designs. We also consider specific applications by applying these design principles and build prototypes to demonstrate the effectiveness of the solutions. In the context of application acceleration, even though all the challenges belong to the two aforementioned dimensions of advanced network protocol design and overcoming application behavior are addressed, application performance can still be limited by the underlying network capability, particularly physical bandwidth. In this work, we study the possibility of speeding up data delivery by eliminating traffic redundancy present in application traffics. Specifically, we first study the traffic redundancy along multiple dimensions using traces obtained from multiple real wireless network deployments. Based on the insights obtained from the analysis, we propose Wireless Memory (WM), a two-ended AP-client solution to effectively exploit traffic redundancy in wireless and mobile environments. Application acceleration can be achieved along two other dimensions: network provision ing and quality of service (QoS). Network provisioning allocates network resources such as physical bandwidth or wireless spectrum, while QoS provides different priority to different applications, users, or data flows. These two dimensions have their respective limitations in the context of application acceleration. In this work, we focus on the two dimensions of overcoming application behavior and Eliminating traffic redundancy to improve application performance. The contribution of this work is as follows. First, we study the problem of application acceleration for wireless and mobile data networks, and we characterize the dimensions along which to address the problem. Second, we identify that application behaviors can significantly affect application performance, and we propose a set of design principles to deal with the behaviors. We also build prototypes to conduct system research. Third, we consider traffic redundancy elimination and propose a wireless memory approach.Ph.D.Committee Chair: Sivakumar, Raghupathy; Committee Member: Ammar, Mostafa; Committee Member: Fekri, Faramarz; Committee Member: Ji, Chuanyi; Committee Member: Ramachandran, Umakishor

    Online Virtual Network Provisioning in Distributed Cloud Computing Data Centers

    Get PDF
    Efficient virtualization methodologies constitute the core of cloud computing data center implementation. Clients are attracted to the cloud model by the ability to scale available resources dynamically and the flexibility in payment options. However, performance hiccups can push them to return to the buy-and-maintain model. Virtualization plays a key role in the synchronous management of the thousands of servers along with clients\u27 data residing on them. To achieve seamless virtualization, cloud providers require a system that performs the function of virtual network mapping. This includes receiving the cloud client requests and allocating computational and network resources in a way that guarantees the quality of service conditions for clients while maximizing the data center resource utilization and providers\u27 revenue. In this thesis, we introduce a comprehensive system to solve the problem of virtual network mapping for a set of connection requests sent by cloud clients. Connections are collected in time intervals called windows. Subsequently, node mapping and link mapping are performed. Different window size selection schemes are introduced and evaluated. Three schemes to prioritize connections are used and their effect is assessed. Moreover, a technique dealing with connections spanning over more than a window is introduced. Simulation results show that the dynamic window size algorithm achieves cloud service providers objectives in terms of generated revenue, served connections ratio, resource utilization and computational overhead. In addition, experimental results show that handling spanning connections independently improves the results for the performance metrics measured. Moreover, in a cloud infrastructure, handling all resources efficiently in their usage, management and energy consumption is challenging. We propose an energy efficient technique for embedding online virtual network requests in cloud data centers. The core focus of this study is to manage energy efficiently in cloud environment. A fixed windowing technique with spanning connections is used. Our algorithm, and a technique for randomly embedding nodes and links are also explained. The results clearly show that the algorithm used in this study generated better results in terms of energy consumption, served connections and revenue generation

    Control Strategies for Improving Cloud Service Robustness

    Get PDF
    This thesis addresses challenges in increasing the robustness of cloud-deployed applications and services to unexpected events and dynamic workloads. Without precautions, hardware failures and unpredictable large traffic variations can quickly degrade the performance of an application due to mismatch between provisioned resources and capacity needs. Similarly, disasters, such as power outages and fire, are unexpected events on larger scale that threatens the integrity of the underlying infrastructure on which an application is deployed.First, the self-adaptive software concept of brownout is extended to replicated cloud applications. By monitoring the performance of each application replica, brownout is able to counteract temporary overload situations by reducing the computational complexity of jobs entering the system. To avoid existing load balancers interfering with the brownout functionality, brownout-aware load balancers are introduced. Simulation experiments show that the proposed load balancers outperform existing load balancers in providing a high quality of service to as many end users as possible. Experiments in a testbed environment further show how a replicated brownout-enabled application is able to maintain high performance during overloads as compared to its non-brownout equivalent.Next, a feedback controller for cloud autoscaling is introduced. Using a novel way of modeling the dynamics of typical cloud application, a mechanism similar to the classical Smith predictor to compensate for delays in reconfiguring resource provisioning is presented. Simulation experiments show that the feedback controller is able to achieve faster control of the response times of a cloud application as compared to a threshold-based controller.Finally, a solution for handling the trade-off between performance and disaster tolerance for geo-replicated cloud applications is introduced. An automated mechanism for differentiating application traffic and replication traffic, and dynamically managing their bandwidth allocations using an MPC controller is presented and evaluated in simulation. Comparisons with commonly used static approaches reveal that the proposed solution in overload situations provides increased flexibility in managing the trade-off between performance and data consistency
    corecore