19 research outputs found

    Understanding PCIe performance for end host networking

    Get PDF
    In recent years, spurred on by the development and availability of programmable NICs, end hosts have increasingly become the enforcement point for core network functions such as load balancing, congestion control, and application specific network offloads. However, implementing custom designs on programmable NICs is not easy: many potential bottlenecks can impact performance. This paper focuses on the performance implication of PCIe, the de-facto I/O interconnect in contemporary servers, when interacting with the host architecture and device drivers. We present a theoretical model for PCIe and pcie-bench, an open-source suite, that allows developers to gain an accurate and deep understanding of the PCIe substrate. Using pcie-bench, we characterize the PCIe subsystem in modern servers. We highlight surprising differences in PCIe implementations, evaluate the undesirable impact of PCIe features such as IOMMUs, and show the practical limits for common network cards operating at 40Gb/s and beyond. Furthermore, through pcie-bench we gained insights which guided software and future hardware architectures for both commercial and research oriented network cards and DMA engines

    Dahu: Improved Data Center Multipath Forwarding

    Full text link

    A hybrid network architecture for modular data centers

    Full text link
    The emergence of the mega data center has resulted in the basic building block of ever larger data centers changing from a rack comprising tens of servers to a self-contained modular shipping container that holds upto a thousand servers. These self-contained modular blocks include networking, power and cooling equipment besides servers. However, provisioning bandwidth between these containers at a large scale is still a significant challenge. Traditional approaches to provisioning bandwidth use electrical packet switches with a scale-up architecture and are often highly oversubscribed. More recent proposals such as those using clos networks promise full bisection bandwidth between servers, albeit at a high cost and power consumption. We present Helios, a hybrid architecture for modular data centers that combines electrical packet switching and optical circuit switching in a single network and dynamically provisions bandwidth between the modular containers on demand. We investigate this design from an architectural standpoint by building a fully functional prototype and explore its implications for data center networks and the challenges that it introduces. Our prototype shows the feasibility of building such a system and achieving high performance at considerably lower cost. Additionally it uncovers several issues that pose new challenges for designing large data center networks and problems that arise when circuits are rapidly reconfigure

    Network Performance Improvements For Web Services : : An End-to-End View

    Full text link
    Modern web services are complex systems with several components that impose stringent performance requirements on the network. The networking subsystem in turn consists of several pieces, such as the wide area and data center networks, different devices, and protocols involved in a user's interaction with a web service. In this dissertation we take a holistic view of the network and improve efficiency and functionality across the stack. We identify three important networking challenges faced by web services in the wide area network, the data center network, and the host network stack, and present solutions. First, web services are dominated by short TCP flows that terminate in as few as 2-3 round trips. Thus, an additional round trip for TCP's connection handshake adds a significant latency overhead. We present TCP Fast Open, a transport protocol enhancement, that enables safe data exchange during TCP's initial handshake, thereby reducing application network latency by a full round trip time. TCP Fast Open uses a security token to verify client IP address ownership, and mitigates the security considerations that arise from allowing data exchange during the handshake. TCP Fast Open is widely deployed and is available as part of the Linux Kernel since version 3.6. Second, provisioning network bandwidth for hundreds of thousands of servers in the data center is expensive. Traditional shortest path based routing protocols are unable to effectively utilize the underlying topology's capacity to maximize network utilization. We present Dahu, a commodity switch design targeted at data centers, that avoids congestion hot-spots by dynamically spreading traffic uniformly across links, and actively leverages non -shortest paths for traffic forwarding. Third, scalable rate limiting is an important primitive for managing server network resources in the data center. Unfortunately, software-based rate limiting suffers from limited accuracy and high CPU overhead at high link speeds, whereas current NICs only support few tens of hardware rate limiters. We present SENIC, a NIC design that natively supports tens of thousands of rate limiters -- 100x to 1000x the number available in NICs today -- to meet the needs of network performance isolation and congestion control in data center

    Hedera: Dynamic flow scheduling for data center networks

    Full text link
    Today’s data centers offer tremendous aggregate bandwidth to clusters of tens of thousands of machines. However, because of limited port densities in even the highest-end switches, data center topologies typically consist of multi-rooted trees with many equal-cost paths between any given pair of hosts. Existing IP multipathing protocols usually rely on per-flow static hashing and can cause substantial bandwidth losses due to longterm collisions. In this paper, we present Hedera, a scalable, dynamic flow scheduling system that adaptively schedules a multi-stage switching fabric to efficiently utilize aggregate network resources. We describe our implementation using commodity switches and unmodified hosts, and show that for a simulated 8,192 host data center, Hedera delivers bisection bandwidth that is 96 % of optimal and up to 113 % better than static load-balancing methods.

    Bullet Genetic Algorithms for Level Control in a Real Time Process

    Full text link
    Measurement of level, temperature, pressure and flow parameters are very vital in all process industries. A combination of a few transducers with a controller, that forms a closed loop system leads to a stable and effective process. Level control of a spherical tank is a complex issue because of the non-linear nature of the tank. The model for such a real time process is identified and validated .The need for improved performance of the process has led to the development of optimal controllers. Genetic Algorithms (GA) is an evolutionary algorithm that is proposed for use in this respect. Determination or tuning of the Proportional-Integral (PI) parameters continues to be important as these parameters have a great influence on the stability and performance of the control system. The methodology and efficiency of proposed method are compared with that of Internal Model Control (IMC) and proves to be better
    corecore