1,382 research outputs found

    Cost Optimization and Load Balancing of Intra and Inter Data Center Networks to Facilitate Cloud Services

    Get PDF
    Title from PDF of title page viewed January 3, 2019Dissertation advisor: Deep MedhiVitaIncludes bibliographical references (pages 127-137}Thesis (Ph.D.)--School of Computing and Engineering. University of Missouri--Kansas City, 2018For cloud enterprise customers that require services on demand, data centers (DC) must allocate and partition data center resources in a dynamic fashion. We consider the problem of allocating data center resources for cloud enterprise customers who require guaranteed services on demand. In particular, a request from an enterprise customer is mapped to a virtual network (VN) class that is allocated both bandwidth and compute resources by connecting it from an entry point of a data center to one or more hosts while there are multiple geographically distributed data centers to choose from. We take a dynamic trafļ¬c engineering approach over multiple time periods in which an energy aware resource reservation model is solved at each review point. In this dissertation, at ļ¬rst for the energy-aware resource reservation problem, we present a mixed-integer linear programming (MILP) formulation (for small-scale problems) and a heuristic approach (for large-scale problems). Our heuristic is fast for solving large-scale problems where the MILP problem becomes difļ¬cult to solve. Through a comprehensive set of studies, we found that a VN class with a low resource requirement has a low blocking even in heavy trafļ¬c, while the VN class with a high resource requirement faces a high service denial. Furthermore, the VN class having randomly distributed resource requirement has a high provisioning cost and blocking compared to the VN class having the same resource requirement for each request although the average resource requirement is same for both these VN classes. We also observe that our approach reduces the maximum energy consumption by about one-sixth at the low arrival rate to by about one-third at the highest arrival rate which also depends on how many different CPU frequency levels a server can run at. Allocation of resources in data centers needs to be done in a dynamic fashion for cloud enterprise customers who require virtualized reservation-oriented services on demand. Due to the spatial diversity of data centers, the cost of using different DCs also varies. In this dissertation, we then propose an allocation scheme to balance the load among these DCs with different cost to minimize the total provisioning cost in a dynamic environment while ensuring that the service level agreements (SLAs) are met. Compared to a benchmark scheme (where all requests are ļ¬rst sent to the cheapest data center), our scheme can decrease the proportional utilization from 24% (for heavy load) to 30% (for normal load) and achieve a signiļ¬cant balance in the cost incurred by individual DCs. Our scheme can also achieve 7.5% reduction in total provisioning cost under certain service level agreement (SLA) in exchange of low increment in blocking. Furthermore, we tested our scheme on 5 DCs to show that our allocation schemes follows the weighted cost proportionally. With the increasing dependency of cloud-based services, data centers have be come a popular platform to satisfy customersā€™ requests. Many large network providers now have their own geographically distributed DCs for cloud services, or have partner ships with third party DC providers to route customersā€™ demand. When end customersā€™ re quests arrive at a Point-of-Presence (PoP) of a large Internet Service Provider, the provider having DCs in multiple geo-locations needs to decide which DC should serve the request depending on the geo-distance, cost of resources in that DC, availability of the requested resource at that DC, and congestion in the path from the customersā€™ location to that DC. Therefore, an optimal connectivity scheme from the ingress PoP to egress DC is required among the PoPs and DCs to minimize the cost of establishing paths between a PoP and a DC while ensuring load balancing in both the link level and DC level. Considering these, we also present a novel mix-integer linear programming (MILP) model for this problem. We show the efļ¬cacy of our model through various performance metrics such as average and maximum link utilization, and average number of links used per path.Introduction -- Literature review -- Model and heuristic for intra DC cost optimization -- Simulation setup and result analysis for intra DC cost optimization -- Load balancing in geo-distributed data centers -- Optimal connectivity between inter DC networks -- Conclusion and future research -- Appendix A. Intra DC optimization model in AMPL -- Appendix B. Optimal connectivity to inter DC network model in AMP

    A Literature Survey on Resource Management Techniques, Issues and Challenges in Cloud Computing

    Get PDF
    Cloud computing is a large scale distributed computing which provides on demand services for clients. Cloud Clients use web browsers, mobile apps, thin clients, or terminal emulators to request and control their cloud resources at any time and anywhere through the network. As many companies are shifting their data to cloud and as many people are being aware of the advantages of storing data to cloud, there is increasing number of cloud computing infrastructure and large amount of data which lead to the complexity management for cloud providers. We surveyed the state-of-the-art resource management techniques for IaaS (infrastructure as a service) in cloud computing. Then we put forward different major issues in the deployment of the cloud infrastructure in order to avoid poor service delivery in cloud computing

    Dynamic Capacity Enhancement using Air Computing: An Earthquake Case

    Full text link
    Earthquakes are one of the most destructive natural disasters harming life and the infrastructure of cities. After an earthquake, functioning communication and computational capacity are crucial for rescue teams and healthcare of victims. Therefore, an earthquake can be investigated for dynamic capacity enhancement in which additional resources are deployed since the surviving portion of the infrastructure may not meet the demand of the users. In this study, we propose a new computation paradigm, air computing, which is the air vehicle assisted next generation edge computing through different air platforms, in order to enhance the capacity of the areas affected by an earthquake. To this end, we put forward a novel paradigm that presents a dynamic, responsive, and high-resolution computation environment by explaining its corresponding components, air layers, and essential advantages. Moreover, we focus on the unmanned aerial vehicle (UAV) deployment problem and apply three different methods including the emergency method, the load balancing method, and the location selection index (LSI) method in which we take the delay requirements of applications into account. To test and compare their performance in terms of the task success rate, we developed an earthquake scenario in which three towns are affected with different severity. The experimental results showed that each method can be beneficial considering the circumstances, and goal of the rescue.Comment: 10 pages, 7 figure

    MACHS: Mitigating the Achilles Heel of the Cloud through High Availability and Performance-aware Solutions

    Get PDF
    Cloud computing is continuously growing as a business model for hosting information and communication technology applications. However, many concerns arise regarding the quality of service (QoS) offered by the cloud. One major challenge is the high availability (HA) of cloud-based applications. The key to achieving availability requirements is to develop an approach that is immune to cloud failures while minimizing the service level agreement (SLA) violations. To this end, this thesis addresses the HA of cloud-based applications from different perspectives. First, the thesis proposes a componentā€™s HA-ware scheduler (CHASE) to manage the deployments of carrier-grade cloud applications while maximizing their HA and satisfying the QoS requirements. Second, a Stochastic Petri Net (SPN) model is proposed to capture the stochastic characteristics of cloud services and quantify the expected availability offered by an application deployment. The SPN model is then associated with an extensible policy-driven cloud scoring system that integrates other cloud challenges (i.e. green and cost concerns) with HA objectives. The proposed HA-aware solutions are extended to include a live virtual machine migration model that provides a trade-off between the migration time and the downtime while maintaining HA objective. Furthermore, the thesis proposes a generic input template for cloud simulators, GITS, to facilitate the creation of cloud scenarios while ensuring reusability, simplicity, and portability. Finally, an availability-aware CloudSim extension, ACE, is proposed. ACE extends CloudSim simulator with failure injection, computational paths, repair, failover, load balancing, and other availability-based modules

    A survey on OFDM-based elastic core optical networking

    Get PDF
    Orthogonal frequency-division multiplexing (OFDM) is a modulation technology that has been widely adopted in many new and emerging broadband wireless and wireline communication systems. Due to its capability to transmit a high-speed data stream using multiple spectral-overlapped lower-speed subcarriers, OFDM technology offers superior advantages of high spectrum efficiency, robustness against inter-carrier and inter-symbol interference, adaptability to server channel conditions, etc. In recent years, there have been intensive studies on optical OFDM (O-OFDM) transmission technologies, and it is considered a promising technology for future ultra-high-speed optical transmission. Based on O-OFDM technology, a novel elastic optical network architecture with immense flexibility and scalability in spectrum allocation and data rate accommodation could be built to support diverse services and the rapid growth of Internet traffic in the future. In this paper, we present a comprehensive survey on OFDM-based elastic optical network technologies, including basic principles of OFDM, O-OFDM technologies, the architectures of OFDM-based elastic core optical networks, and related key enabling technologies. The main advantages and issues of OFDM-based elastic core optical networks that are under research are also discussed

    REACT: A Solidarity-based Elastic Service Resource Reallocation Strategy for Multi-access Edge Computing

    Get PDF
    The Multi-access Edge Computing (MEC) paradigm promises to enhance network flexibility and scalability through resource virtualization. MEC allows telecom operators to fulfill the stringent and heterogeneous requirements of 5G applications via service deployment at the edge of the mobile network. However, current solutions to support MEC struggle to provide resource elasticity since MEC infrastructures have limited resources. The coexistence of many heterogeneous services on the distributed MEC infrastructure makes the resource scarcity problem even more challenging than it already is in traditional networks. Services need distinct resource provisioning patterns due to their diverse requirements, and we may not assume an extensive MEC infrastructure that can accommodate an arbitrary number of services. To address these aspects, we present REACT: a MEC-suppoRted sElfadaptive elAstiCiTy mechanism that leverages resource provisioning among different services running on a shared MEC environment. REACT adopts an adaptive and solidarity-based strategy to redistribute resources from over-provisioned services to under-provisioned services in MEC environments. REACT is an alternative strategy to avoid service migration due to resource scarcity. Real testbed results show that REACT outperforms Kubernetesā€™ elasticity strategy by accomplishing up to 18.88% more elasticity events, reducing service outages by up to 95.1%, reducing elasticity attempts by up to 95.36%, and reducing over-provisioned resources by up to 33.88%, 38.41%, and 73% for CPU cycles, RAM and bandwidth resources, respectively. Finally, REACT reduces response time by up to 15.5%
    • ā€¦
    corecore