1,819 research outputs found

    DCCast: Efficient Point to Multipoint Transfers Across Datacenters

    Full text link
    Using multiple datacenters allows for higher availability, load balancing and reduced latency to customers of cloud services. To distribute multiple copies of data, cloud providers depend on inter-datacenter WANs that ought to be used efficiently considering their limited capacity and the ever-increasing data demands. In this paper, we focus on applications that transfer objects from one datacenter to several datacenters over dedicated inter-datacenter networks. We present DCCast, a centralized Point to Multi-Point (P2MP) algorithm that uses forwarding trees to efficiently deliver an object from a source datacenter to required destination datacenters. With low computational overhead, DCCast selects forwarding trees that minimize bandwidth usage and balance load across all links. With simulation experiments on Google's GScale network, we show that DCCast can reduce total bandwidth usage and tail Transfer Completion Times (TCT) by up to 50%50\% compared to delivering the same objects via independent point-to-point (P2P) transfers.Comment: 9th USENIX Workshop on Hot Topics in Cloud Computing, https://www.usenix.org/conference/hotcloud17/program/presentation/noormohammadpou

    Joint Elastic Cloud and Virtual Network Framework for Application Performance-cost Optimization

    Get PDF
    International audienceCloud computing infrastructures are providing resources on demand for tackling the needs of large-scale distributed applications. To adapt to the diversity of cloud infras- tructures and usage, new operation tools and models are needed. Estimating the amount of resources consumed by each application in particular is a difficult problem, both for end users who aim at minimizing their costs and infrastructure providers who aim at control- ling their resources allocation. Furthermore, network provision is generally not controlled on clouds. This paper describes a framework automating cloud resources allocation, deploy- ment and application execution control. It is based on a cost estimation model taking into account both virtual network and nodes managed by the cloud. The flexible provisioning of network resources permits the optimization of applications performance and infrastructure cost reduction. Four resource allocation strategies relying on the expertise that can be cap- tured in workflow-based applications are considered. Results of these strategies are confined virtual infrastructure descriptions that are interpreted by the HIPerNet engine responsible for allocating, reserving and configuring physical resources. The evaluation of this framework was carried out on the Aladdin/Grid'5000 testbed using a real application from the area of medical image analysis

    Improvement of indoor VLC network downlink scheduling and resource allocation

    Get PDF
    Indoor visible light communications (VLC) combines illumination and communication by utilizing the high-modulation-speed of LEDs. VLC is anticipated to be complementary to radio frequency communications and an important part of next generation heterogeneous networks. In order to make the maximum use of VLC technology in a networking environment, we need to expand existing research from studies of traditional point-to-point links to encompass scheduling and resource allocation related to multi-user scenarios. This work aims to maximize the downlink throughput of an indoor VLC network, while taking both user fairness and time latency into consideration. Inter-user interference is eliminated by appropriately allocating LEDs to users with the aid of graph theory. A three-term priority factor model is derived and is shown to improve the throughput performance of the network scheduling scheme over those previously reported. Simulations of VLC downlink scheduling have been performed under proportional fairness scheduling principles where our newly formulated priority factor model has been applied. The downlink throughput is improved by 19.6% compared to previous two-term priority models, while achieving similar fairness and latency performance. When the number of users grows larger, the three-term priority model indicates an improvement in Fairness performance compared to two-term priority model scheduling
    • …
    corecore