1,012 research outputs found

    A disaster-resilient multi-content optical datacenter network architecture

    Get PDF
    Cloud services based on datacenter networks are becoming very important. Optical networks are well suited to meet the demands set by the high volume of traffic between datacenters, given their high bandwidth and low-latency characteristics. In such networks, path protection against network failures is generally ensured by providing a backup path to the same destination, which is link-disjoint to the primary path. This protection fails to protect against disasters covering an area which disrupts both primary and backup resources. Also, content/service protection is a fundamental problem in datacenter networks, as the failure of a single datacenter should not cause the disappearance of a specific content/service from the network. Content placement, routing and protection of paths and content are closely related to one another, so the interaction among these should be studied together. In this work, we propose an integrated ILP formulation to design an optical datacenter network, which solves all the above-mentioned problems simultaneously. We show that our disaster protection scheme exploiting anycasting provides more protection, but uses less capacity, than dedicated single-link protection. We also show that a reasonable number of datacenters and selective content replicas with intelligent network design can provide survivability to disasters while supporting user demands

    Fastpass: A Centralized “Zero-Queue” Datacenter Network

    Get PDF
    An ideal datacenter network should provide several properties, including low median and tail latency, high utilization (throughput), fair allocation of network resources between users or applications, deadline-aware scheduling, and congestion (loss) avoidance. Current datacenter networks inherit the principles that went into the design of the Internet, where packet transmission and path selection decisions are distributed among the endpoints and routers. Instead, we propose that each sender should delegate control—to a centralized arbiter—of when each packet should be transmitted and what path it should follow. This paper describes Fastpass, a datacenter network architecture built using this principle. Fastpass incorporates two fast algorithms: the first determines the time at which each packet should be transmitted, while the second determines the path to use for that packet. In addition, Fastpass uses an efficient protocol between the endpoints and the arbiter and an arbiter replication strategy for fault-tolerant failover. We deployed and evaluated Fastpass in a portion of Facebook’s datacenter network. Our results show that Fastpass achieves high throughput comparable to current networks at a 240 reduction is queue lengths (4.35 Mbytes reducing to 18 Kbytes), achieves much fairer and consistent flow throughputs than the baseline TCP (5200 reduction in the standard deviation of per-flow throughput with five concurrent connections), scalability from 1 to 8 cores in the arbiter implementation with the ability to schedule 2.21 Terabits/s of traffic in software on eight cores, and a 2.5 reduction in the number of TCP retransmissions in a latency-sensitive service at Facebook.National Science Foundation (U.S.) (grant IIS-1065219)Irwin Mark Jacobs and Joan Klein Jacobs Presidential FellowshipHertz Foundation (Fellowship
    • …
    corecore