4,063 research outputs found

    InterCloud: Utility-Oriented Federation of Cloud Computing Environments for Scaling of Application Services

    Full text link
    Cloud computing providers have setup several data centers at different geographical locations over the Internet in order to optimally serve needs of their customers around the world. However, existing systems do not support mechanisms and policies for dynamically coordinating load distribution among different Cloud-based data centers in order to determine optimal location for hosting application services to achieve reasonable QoS levels. Further, the Cloud computing providers are unable to predict geographic distribution of users consuming their services, hence the load coordination must happen automatically, and distribution of services must change in response to changes in the load. To counter this problem, we advocate creation of federated Cloud computing environment (InterCloud) that facilitates just-in-time, opportunistic, and scalable provisioning of application services, consistently achieving QoS targets under variable workload, resource and network conditions. The overall goal is to create a computing environment that supports dynamic expansion or contraction of capabilities (VMs, services, storage, and database) for handling sudden variations in service demands. This paper presents vision, challenges, and architectural elements of InterCloud for utility-oriented federation of Cloud computing environments. The proposed InterCloud environment supports scaling of applications across multiple vendor clouds. We have validated our approach by conducting a set of rigorous performance evaluation study using the CloudSim toolkit. The results demonstrate that federated Cloud computing model has immense potential as it offers significant performance gains as regards to response time and cost saving under dynamic workload scenarios.Comment: 20 pages, 4 figures, 3 tables, conference pape

    Software-Defined Cloud Computing: Architectural Elements and Open Challenges

    Full text link
    The variety of existing cloud services creates a challenge for service providers to enforce reasonable Software Level Agreements (SLA) stating the Quality of Service (QoS) and penalties in case QoS is not achieved. To avoid such penalties at the same time that the infrastructure operates with minimum energy and resource wastage, constant monitoring and adaptation of the infrastructure is needed. We refer to Software-Defined Cloud Computing, or simply Software-Defined Clouds (SDC), as an approach for automating the process of optimal cloud configuration by extending virtualization concept to all resources in a data center. An SDC enables easy reconfiguration and adaptation of physical resources in a cloud infrastructure, to better accommodate the demand on QoS through a software that can describe and manage various aspects comprising the cloud environment. In this paper, we present an architecture for SDCs on data centers with emphasis on mobile cloud applications. We present an evaluation, showcasing the potential of SDC in two use cases-QoS-aware bandwidth allocation and bandwidth-aware, energy-efficient VM placement-and discuss the research challenges and opportunities in this emerging area.Comment: Keynote Paper, 3rd International Conference on Advances in Computing, Communications and Informatics (ICACCI 2014), September 24-27, 2014, Delhi, Indi

    Algorithms for advance bandwidth reservation in media production networks

    Get PDF
    Media production generally requires many geographically distributed actors (e.g., production houses, broadcasters, advertisers) to exchange huge amounts of raw video and audio data. Traditional distribution techniques, such as dedicated point-to-point optical links, are highly inefficient in terms of installation time and cost. To improve efficiency, shared media production networks that connect all involved actors over a large geographical area, are currently being deployed. The traffic in such networks is often predictable, as the timing and bandwidth requirements of data transfers are generally known hours or even days in advance. As such, the use of advance bandwidth reservation (AR) can greatly increase resource utilization and cost efficiency. In this paper, we propose an Integer Linear Programming formulation of the bandwidth scheduling problem, which takes into account the specific characteristics of media production networks, is presented. Two novel optimization algorithms based on this model are thoroughly evaluated and compared by means of in-depth simulation results

    Virtual Machines Embedding for Cloud PON AWGR and Server Based Data Centres

    Full text link
    In this study, we investigate the embedding of various cloud applications in PON AWGR and Server Based Data Centres

    A Minimum-Cost Flow Model for Workload Optimization on Cloud Infrastructure

    Full text link
    Recent technology advancements in the areas of compute, storage and networking, along with the increased demand for organizations to cut costs while remaining responsive to increasing service demands have led to the growth in the adoption of cloud computing services. Cloud services provide the promise of improved agility, resiliency, scalability and a lowered Total Cost of Ownership (TCO). This research introduces a framework for minimizing cost and maximizing resource utilization by using an Integer Linear Programming (ILP) approach to optimize the assignment of workloads to servers on Amazon Web Services (AWS) cloud infrastructure. The model is based on the classical minimum-cost flow model, known as the assignment model.Comment: 2017 IEEE 10th International Conference on Cloud Computin

    Energy Efficient Cloud Networks

    Get PDF
    Cloud computing is expected to be a major factor that will dominate the future Internet service model. This paper summarizes our work on energy efficiency for cloud networks. We develop a framework for studying the energy efficiency of four cloud services in IP over WDM networks: cloud content delivery, storage as a service (StaaS), and virtual machines (VMS) placement for processing applications and infrastructure as a service (IaaS).Our approach is based on the co-optimization of both external network related factors such as whether to geographically centralize or distribute the clouds, the influence of users’ demand distribution, content popularity, access frequency and renewable energy availability and internal capability factors such as the number of servers, switches and routers as well as the amount of storage demanded in each cloud. Our investigation of the different energy efficient approaches is backed with Mixed Integer Linear Programming (MILP) models and real time heuristic

    Energy Efficient Resource Allocation for Virtual Network Services with Dynamic Workload in Cloud Data Centers

    Get PDF
    Title from PDF of title page, viewed on March 21, 2016Dissertation advisor: Baek-Young ChoiVitaIncludes bibliographical references (pages 126-143)Thesis (Ph.D.)--School of Computing and Engineering. University of Missouri--Kansas City, 2016With the rapid proliferation of cloud computing, more and more network services and applications are deployed on cloud data centers. Their energy consumption and green house gas emissions have significantly increased. Some efforts have been made to control and lower energy consumption of data centers such as, proportional energy consuming hardware, dynamic provisioning, and virtualization machine techniques. However, it is still common that many servers and network resources are often underutilized, and idle servers spend a large portion of their peak power consumption. Network virtualization and resource sharing have been employed to improve energy efficiency of data centers by aggregating workload to a few physical nodes and switch the idle nodes to sleep mode. Especially, with the advent of live migration, a virtual node can be moved from one physical node to another physical node without service disrup tion. It is possible to save more energy by shrinking virtual nodes to a small set of physical nodes and turning the idle nodes to sleep mode when the service workload is low, and expanding virtual nodes to a large set of physical nodes to satisfy QoS requirements when the service workload is high. When the service provider explicates the desired virtual network including a specific topology, and a set of virtual nodes with certain resource demands, the infrastructure provider computes how the given virtual network is embedded to its operated data centers with minimum energy consumption. When the service provider only gives some description about the network service and the desired QoS requirements, the infrastructure provider has more freedom on how to allocate resources for the network service. For the first problem, we consider the evolving workload of the virtual networks or virtual applications and residual resources in data centers, and build a novel model of energy efficient virtual network embedding (EE-VNE) in order to minimize energy usage in the physical network consists of multiple data centers. In this model, both operation cost for executing network services’ task and migration cost for the live migrations of virtual nodes are counted toward the total energy consumption. In addition, rather than random generated physical network topology, we use practical assumption about physical network topology in our model. Due to the NP-hardness of the proposed model, we develop a heuristic algorithm for virtual network scheduling and mapping. In doing so, we specifically take the expected energy consumption at different times, virtual network operation and future migration costs, and a data center architecture into consideration. Our extensive evaluation results showthatouralgorithmcouldreduceenergyconsumptionupto40%andtakeuptoa57% higher number of virtual network requests over other existing virtual mapping schemes. However, through comparison with CPLEX based exact algorithm, we identify that there is still a gap between the heuristic solution and the optimal solution. Therefore, after investigation other solutions, we convert the origin EE-VNE problem to an Ant Colony Optimization (ACO) problem by building the construction model and presenting the transition probability formula. Then, ACO based algorithm has been adapted to solve the ACO-EE-VNE problem. In addition, we reduce the space complexity of ACO-EE VNE by developing a novel way to track and update the pheromone. For the second problem, we design a framework to dynamically allocate resources for a network service by employing container based virtual nodes. In the framework,each network service would have a pallet container and a set of execution containers. The pal let container requests resource based on certain strategy, creates execution containers with assigned resources and manage the life cycle of the containers; while the execution containers execute the assigned job for the network service. Formulations are presented to optimize resource usage efficiency and save energy consumption for network services with dynamic workload, and a heuristic algorithm is proposed to solve the optimization problem. Our numerical results show that container based resource allocation provide more flexible and saves more cost than virtual service deployment with fixed virtual machines and demands. In addition, we study the content distribution problem with joint optimization goal and varied size of contents in cloud storage. Previous research on content distribution mainly focuses on reducing latency experienced by content customers. A few recent studies address the issue of bandwidth usage in CDNs, as the bandwidth consumption is an important issue due to its relevance to the cost of content providers. However, few researches consider both bandwidth consumption and delay performance for the content providers that use cloud storages with limited budgets, which is the focus of this study. We develop an efficient light-weight approximation algorithm toward the joint optimization problem of content placement. We also conduct the analysis of its theoretical complexities. The performance bound of the proposed approximation algorithm exhibits a much better worst case than those in previous studies. We further extend the approximate algorithm into a distributed version that allows it to promptly react to dynamic changes in users’ interests. The extensive results from both simulations and Planetlab experiments exhibit that the performance is near optimal for most of the practical conditions.Introduction -- Related work -- Energy efficient virtual network embedding for green data centers using data center topology and future migration -- Ant colony optimization based energy efficient virtual network embedding -- Energy aware container based resource allocation for virtual services in green data centers -- Achieving optimal content delivery using cloud storage -- Conclusions and future wor
    corecore