598 research outputs found

    Scalable and Cost Efficient Algorithms for Virtual CDN Migration

    Full text link
    Virtual Content Delivery Network (vCDN) migration is necessary to optimize the use of resources and improve the performance of the overall SDN/NFV-based CDN function in terms of network operator cost reduction and high streaming quality. It requires intelligent and enticed joint SDN/NFV migration algorithms due to the evident huge amount of traffic to be delivered to end customers of the network. In this paper, two approaches for finding the optimal and near optimal path placement(s) and vCDN migration(s) are proposed (OPAC and HPAC). Moreover, several scenarios are considered to quantify the OPAC and HPAC behaviors and to compare their efficiency in terms of migration cost, migration time, vCDN replication number, and other cost factors. Then, they are implemented and evaluated under different network scales. Finally, the proposed algorithms are integrated in an SDN/NFV framework. Index Terms: vCDN; SDN/NFV Optimization; Migration Algorithms; Scalability Algorithms.Comment: 9 pages, 11 figures, 4 tableaux, conference Local Computer Networks (LCN), class

    InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services

    Full text link
    Cloud computing providers have setup several data centers at different geographical locations over the Internet in order to optimally serve needs of their customers around the world. However, existing systems do not support mechanisms and policies for dynamically coordinating load distribution among different Cloud-based data centers in order to determine optimal location for hosting application services to achieve reasonable QoS levels. Further, the Cloud computing providers are unable to predict geographic distribution of users consuming their services, hence the load coordination must happen automatically, and distribution of services must change in response to changes in the load. To counter this problem, we advocate creation of federated Cloud computing environment (InterCloud) that facilitates just-in-time, opportunistic, and scalable provisioning of application services, consistently achieving QoS targets under variable workload, resource and network conditions. The overall goal is to create a computing environment that supports dynamic expansion or contraction of capabilities (VMs, services, storage, and database) for handling sudden variations in service demands. This paper presents vision, challenges, and architectural elements of InterCloud for utility-oriented federation of Cloud computing environments. The proposed InterCloud environment supports scaling of applications across multiple vendor clouds. We have validated our approach by conducting a set of rigorous performance evaluation study using the CloudSim toolkit. The results demonstrate that federated Cloud computing model has immense potential as it offers significant performance gains as regards to response time and cost saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape

    On the Load Balancing of Edge Computing Resources for On-Line Video Delivery

    Get PDF
    Online video broadcasting platforms are distributed, complex, cloud oriented, scalable, micro-service-based systems that are intended to provide over-the-top and live content to audience in scattered geographic locations. Due to the nature of cloud VM hosting costs, the subscribers are usually served under limited resources in order to minimize delivery budget. However, operations including transcoding require high-computational capacity and any disturbance in supplying requested demand might result in quality of experience (QoE) deterioration. For any online delivery deployment, understanding user's QoE plays a crucial role for rebalancing cloud resources. In this paper, a methodology for estimating QoE is provided for a scalable cloud-based online video platform. The model will provide an adeptness guideline regarding limited cloud resources and relate computational capacity, memory, transcoding and throughput capability, and finally latency competence of the cloud service to QoE. Scalability and efficiency of the system are optimized through reckoning sufficient number of VMs and containers to satisfy the user requests even on peak demand durations with minimum number of VMs. Both horizontal and vertical scaling strategies (including VM migration) are modeled to cover up availability and reliability of intermediate and edge content delivery network cache nodes

    Towards Cost-effective On-demand continuous Media Service: A Peer-to- Peer Approach

    Get PDF
    To overcome the limited bandwidth of streaming servers, Content Distribution Networks (CDNs) are deployed on the edge of the Internet. A large number of such servers have to be installed to make the whole system scalable, making CDN a very expensive way to distribute media. The primary concern of this research is to find an inexpensive way to alleviate the traffic load for media streaming on the servers in a continuous media service infrastructure. Our approach to solve the above problem is motivated by the emerging concept of peer-to-peer computing. Specifically, we let clients that obtained a media object act as streaming servers for following requests to that media object. Unlike the server/client scheme, peers are heterogeneous in the storage capacity and out-bound bandwidth they can contribute. Secondly, peers are heterogeneous in the duration of their commitment to the community. In our research, we identified the following problems in the context of peer-to-peer streaming: 1) How does the come-and-go behaviors of peers affect the system performance? 2). How do we manage the limited resources contributed by each peer? 3). The design of a peer-to-peer streaming protocol that handles peer failure. Solutions to these problems are described and analyzed. 1

    Analysis of current middleware used in peer-to-peer and grid implementations for enhancement by catallactic mechanisms

    Get PDF
    This deliverable describes the work done in task 3.1, Middleware analysis: Analysis of current middleware used in peer-to-peer and grid implementations for enhancement by catallactic mechanisms from work package 3, Middleware Implementation. The document is divided in four parts: The introduction with application scenarios and middleware requirements, Catnets middleware architecture, evaluation of existing middleware toolkits, and conclusions. -- Die Arbeit definiert Anforderungen an Grid und Peer-to-Peer Middleware Architekturen und analysiert diese auf ihre Eignung fĂĽr die prototypische Umsetzung der Katallaxie. Eine Middleware-Architektur fĂĽr die Umsetzung der Katallaxie in Application Layer Netzwerken wird vorgestellt.Grid Computing

    A Taxonomy of Data Grids for Distributed Data Sharing, Management and Processing

    Full text link
    Data Grids have been adopted as the platform for scientific communities that need to share, access, transport, process and manage large data collections distributed worldwide. They combine high-end computing technologies with high-performance networking and wide-area storage management techniques. In this paper, we discuss the key concepts behind Data Grids and compare them with other data sharing and distribution paradigms such as content delivery networks, peer-to-peer networks and distributed databases. We then provide comprehensive taxonomies that cover various aspects of architecture, data transportation, data replication and resource allocation and scheduling. Finally, we map the proposed taxonomy to various Data Grid systems not only to validate the taxonomy but also to identify areas for future exploration. Through this taxonomy, we aim to categorise existing systems to better understand their goals and their methodology. This would help evaluate their applicability for solving similar problems. This taxonomy also provides a "gap analysis" of this area through which researchers can potentially identify new issues for investigation. Finally, we hope that the proposed taxonomy and mapping also helps to provide an easy way for new practitioners to understand this complex area of research.Comment: 46 pages, 16 figures, Technical Repor

    A service-oriented hybrid access network and clouds architecture

    Get PDF
    Many telecom operators are deploying their own cloud infrastructure with the two-fold objective of providing cloud services to their customers and enabling network function virtualization. In this article we present an architecture we call SHINE, which focuses on orchestrating cloud with heterogeneous access and core networks. In this architecture intra and inter DC connectivity is dynamically controlled, maximizing the overall performance in terms of throughput and latency while minimizing total costs. The main building blocks are: a future-proof network architecture that can scale to offer potentially unlimited bandwidth based on an active remote node (ARN) to interface end-users and the core network; an innovative distributed DC architecture consisting of micro-DCs placed in selected core locations to accelerate content delivery, reducing core network traffic, and ensuring very low latency; and dynamic orchestration of the distributed DC and access and core network segments. SHINE will provide unprecedented quality of experience, greatly reducing costs by coordinating network and cloud and facilitating service chaining by virtualizing network functions.Peer ReviewedPostprint (author’s final draft

    Cloudbus Toolkit for Market-Oriented Cloud Computing

    Full text link
    This keynote paper: (1) presents the 21st century vision of computing and identifies various IT paradigms promising to deliver computing as a utility; (2) defines the architecture for creating market-oriented Clouds and computing atmosphere by leveraging technologies such as virtual machines; (3) provides thoughts on market-based resource management strategies that encompass both customer-driven service management and computational risk management to sustain SLA-oriented resource allocation; (4) presents the work carried out as part of our new Cloud Computing initiative, called Cloudbus: (i) Aneka, a Platform as a Service software system containing SDK (Software Development Kit) for construction of Cloud applications and deployment on private or public Clouds, in addition to supporting market-oriented resource management; (ii) internetworking of Clouds for dynamic creation of federated computing environments for scaling of elastic applications; (iii) creation of 3rd party Cloud brokering services for building content delivery networks and e-Science applications and their deployment on capabilities of IaaS providers such as Amazon along with Grid mashups; (iv) CloudSim supporting modelling and simulation of Clouds for performance studies; (v) Energy Efficient Resource Allocation Mechanisms and Techniques for creation and management of Green Clouds; and (vi) pathways for future research.Comment: 21 pages, 6 figures, 2 tables, Conference pape
    • …
    corecore