4 research outputs found

    Machine Learning Models for Live Migration Metrics Prediction

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (์„์‚ฌ)-- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2019. 2. Egger, Bernhard.์˜ค๋Š˜๋‚  ๋ฐ์ดํ„ฐ ์„ผํ„ฐ์—์„œ ๊ฐ€์ƒ๋จธ์‹ ์˜ ๋ผ์ด๋ธŒ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ๊ธฐ์ˆ ์€ ๋งค์šฐ ์ค‘์š”ํ•˜๊ฒŒ ์‚ฌ์šฉ๋œ๋‹ค. ํ˜„์กดํ•˜๋Š” ๋ฐ์ดํ„ฐ ์„ผํ„ฐ ๊ด€๋ฆฌ ํ”„๋ ˆ์ž„์›Œํฌ์—์„œ๋Š” ๋ณต์žกํ•œ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ด์šฉํ•˜์—ฌ ์–ธ์ œ, ์–ด๋””์„œ, ์–ด๋””๋กœ ๊ฐ€์ƒ๋จธ์‹ ์˜ ๋งˆ์ด๊ทธ๋ ˆ์…˜์„ ์‹คํ–‰ํ• ์ง€๋ฅผ ๊ฒฐ์ •ํ•œ๋‹ค. ํ•˜์ง€๋งŒ ์–ด๋–ค ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ๋ฐฉ๋ฒ•์„ ์‚ฌ์šฉํ•˜๋Š”์ง€์— ๋”ฐ๋ผ์„œ ์„ฑ๋Šฅ์ด ํฌ๊ฒŒ ์ฐจ์ด๊ฐ€ ๋‚  ์ˆ˜ ์žˆ์Œ์—๋„ ๋ถˆ๊ตฌํ•˜๊ณ  ์ด์— ๋Œ€ํ•œ ๋…ผ์˜๋Š” ์ฃผ์š”ํ•˜๊ฒŒ ๋‹ค๋ค„์ง€์ง€ ์•Š์•˜๋‹ค. ์ด๋Ÿฌํ•œ ์„ฑ๋Šฅ์˜ ์ฐจ์ด๋Š” ๋ผ์ด๋ธŒ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ์ฐจ์ด๋‚˜ ๊ฐ€์ƒ๋จธ์‹ ์— ํ• ๋‹น๋œ ์›Œํฌ๋กœ๋“œ์˜ ์–‘์˜ ์ฐจ์ด ๊ทธ๋ฆฌ๊ณ  ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜์„ ํ•˜๋Š” ๊ณณ๊ณผ ๋ชฉ์  host์˜ ์ƒํƒœ ์ฐจ์ด์— ์˜ํ•˜์—ฌ ์ผ์–ด๋‚œ๋‹ค. ๋น ๋ฅด๊ณ  ์ •ํ™•ํ•˜๊ฒŒ ์˜ฌ๋ฐ”๋ฅธ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ๋ฐฉ๋ฒ•์„ ์ •ํ•˜๋Š” ๊ฒƒ์€ ํ•„์ˆ˜์ ์ธ ๊ณผ์ œ์ด๋‹ค. ์ด๋Ÿฌํ•œ ๊ณผ์ œ๋ฅผ performance model์„ ์ด์šฉํ•˜์—ฌ ํ•ด๊ฒฐํ•  ๊ฒƒ์ด๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š”, ๊ฐ€์ƒ๋จธ์‹ ์˜ ๋ผ์ด๋ธŒ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ์„ฑ๋Šฅ์„ ์˜ˆ์ธกํ•˜๋Š” ์—ฌ๋Ÿฌ ๋จธ์‹  ๋Ÿฌ๋‹ ๋ชจ๋ธ์„ ์ œ์‹œํ•œ๋‹ค. ์—ฌ๊ธฐ์„œ 12๊ฐœ์˜ ์„œ๋กœ ๋‹ค๋ฅธ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์— ๋Œ€ํ•ด 7๊ฐ€์ง€์˜ ๋‹ค๋ฅธ metric๋“ค์„ ์˜ˆ์ธกํ•œ๋‹ค. ์ด ๋ชจ๋ธ์€ ๊ธฐ์กด ์—ฐ๊ตฌ์— ๋น„ํ•ด ํ›จ์”ฌ ์ •ํ™•ํ•œ ์˜ˆ์ธก์„ ์„ฑ๊ณตํ•˜์˜€๋‹ค. ๊ฐ๊ฐ์˜ target metric๊ณผ ์—ฌ๋Ÿฌ ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์— ๋Œ€ํ•˜์—ฌ input feature evaluation์„ ์ˆ˜ํ–‰ํ•˜์˜€๊ณ  ๊ฐ๊ฐ์˜ ํŠน์„ฑ์— ๋งž๋Š” ๋ชจ๋ธ์„ ๋งŒ๋“ค์–ด 84๊ฐœ์˜ ์„œ๋กœ๋‹ค๋ฅธ ๋จธ์‹  ๋Ÿฌ๋‹ ๋ชจ๋ธ๋“ค์„ ํ›ˆ๋ จ์‹œ์ผฐ๋‹ค. ์ด๋Ÿฌํ•œ ๋ชจ๋ธ๋“ค์€ ์‹ค์ œ ๋ผ์ด๋ธŒ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ํ”„๋ ˆ์ž„์›Œํฌ์— ์‰ฝ๊ฒŒ ์ ์šฉ ๊ฐ€๋Šฅํ•˜๋‹ค. ๊ฐ๊ฐ์˜ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์— ๋Œ€ํ•˜์—ฌ target metric ์˜ˆ์ธก์„ ์‚ฌ์šฉํ•จ์œผ๋กœ์จ ์˜ฌ๋ฐ”๋ฅธ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜ ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์‰ฝ๊ฒŒ ๊ฒฐ์ •ํ•  ์ˆ˜ ์žˆ๊ณ  ์ด๋Š” ๊ฒฐ๊ณผ์ ์œผ๋กœ ๋‹ค์šดํƒ€์ž„๊ณผ ๋งˆ์ด๊ทธ๋ ˆ์ด์…˜์— ์†Œ์š”๋˜๋Š” ์ด ์‹œ๊ฐ„์˜ ๊ฐ์†Œ ํšจ๊ณผ๋ฅผ ๋ณผ ์ˆ˜ ์žˆ๋‹ค.Live migration of Virtual Machines (VMs) is an important technique in today's data centers. In existing data center management frameworks, complex algorithms are used to determine when, where, and to which host a migration of a VM is to be performed. However, very little attention is paid to the selection of the right migration technique depending on which the migration performance can vary greatly. This performance fluctuation is caused by the different live migration algorithms, the different workloads that each VM is executing, and the state of the destination and the source host. Choosing the right migration technique is a crucial task that has to be made quickly and precisely. Therefore, a performance model is the best and the right candidate for such a task. In this thesis, we propose various machine learning models for predicting live migration metrics of virtual machines. We predict seven different metrics for twelve distinct migration algorithms. Our models achieve a much higher accuracy compared to existing work. For each target metric and algorithm, an input feature evaluation is conducted and a strictly specific model is generated, leading to 84 different trained machine learning models. These models can easily be integrated into a live migration framework. Using the target metric predictions for each migration algorithm, a framework can easily choose the right migration algorithm, which can lead to downtime and total migration time reduction and less service-level agreement violations.Abstract Contents List of Figures List of Tables Chapter 1 Introduction and Motivation Chapter 2 Background 2.1 Virtualization 2.2 Live Migration 2.3 SLA and SLO 2.4 Live Migration Techniques 2.4.1 Pre-copy (PRE) 2.4.2 Post-copy (POST) 2.4.3 Hybrid Migration Techniques 2.5 Live Migration Performance Metrics 2.6 Artificial Neural Networks 2.6.1 Feedforward Neural Network (FNN) 2.6.2 Deep Neural Network (DNN) 2.6.3 Convolution Neural Network (CNN) Chapter 3 Related Work Chapter 4 Overview and Design Chapter 5 Implementation 5.1 Deep Neural Network design 5.2 Convolutional Neural Network design Chapter 6 Evaluation metrics 6.1 Geometric Mean Absolute Error (GMAE) 6.2 Geometric Mean Relative Error (GMRE) 6.3 Mean Absolute Error (MAE) 6.4 Weighted Absolute Percentage Error (WAPE) Chapter 7 Results 7.1 Deep Neural Network 7.2 SVR with bagging 7.3 DNN vs. SVR comparison 7.4 Overhead Chapter 8 Conclusion and Future Work 8.1 Conclusion 8.2 Future Work AppendicesMaste

    Scalable Network Design and Management with Decentralized Software-defined Networking

    Get PDF
    Network softwarization is among the most significant innovations of computer networks in the last few decades. The lack of uniform and programmable interfaces for network management led to the design of OpenFlow protocol for the university campuses and enterprise networks. This breakthrough coupled with other similar efforts led to an emergence of two complementary but independent paradigms called software-defined networking (SDN) and network function virtualization (NFV). As of this writing, these paradigms are becoming the de-facto norms of wired and wireless networks alike. This dissertation mainly addresses the scalability aspect of SDN for multiple network types. Although centralized control and separation of control and data planes play a pivotal role for ease of network management, these concepts bring in many challenges as well. Scalability is among the most crucial challenges due to the unprecedented growth of computer networks in the past few years. Therefore, we strive to grapple with this problem in diverse networking scenarios and propose novel solutions by harnessing capabilities provided by SDN and other related technologies. Specifically, we present the techniques to deploy SDN at the Internet scale and to extend the concepts of softwarization for mobile access networks and vehicular networks. Multiple optimizations are employed to mitigate latency and other overheads that contribute to achieve performance gains. Additionally, by taking care of sparse connectivity and high mobility, the intrinsic constraints of centralization for wireless ad-hoc networks are addressed in a systematic manner. The state-of-the-art virtualization techniques are coupled with cloud computing methods to exploit the potential of softwarization in general and SDN in particular. Finally, by tapping into the capabilities of machine learning techniques, an SDN-based solution is proposed that inches closer towards the longstanding goal of self-driving networks. Extensive experiments performed on a large-scale testbed corroborates effectiveness of our approaches

    Colocation-aware Resource Management for Distributed Parallel Applications in Consolidated Clusters

    Get PDF
    Department of Computer Science and EngineeringConsolidated clusters, which run various distributed parallel applications such as big data frameworks, machine learning applications, and scienti???c applications to solve complex problems in wide range of fields, are already used commonly. Resource providers allow various applications with different characteristics to execute together to efficiently utilize their resources. There are some important issues about scheduling applications to resources. When applications share the same resources, interference between them affects their performance. The performance of applications can be improved or degraded depending on which resources are used to execute them based on various characteristics of applications and resources. Characteristics and resource requirements of applications can constrain their placement, and these constraints can be extended to constraints between applications. These issues should be considered to manage resource e???ciently and improve the performance of applications. In this thesis, we study how to manage resources e???ciently while scheduling distributed parallel applications in consolidated clusters. First, we present a holistic VM placement technique for distributed parallel applications in heterogeneous virtual cluster, aiming to maximize the e???ciency of the cluster and consequently reduce cost for service providers and users. We analyze the e???ect of heterogeneity of resource, di???erent VM con???gurations, and interference between VMs on the performance of distributed parallel applications and propose a placement technique that uses a machine learning algorithm to estimate the runtime of a distributed parallel application. Second, we present a two-level scheduling algorithms, which distribute applications to platforms then map tasks to each node. we analyze the platform and co-runner a???nities of looselycoupled applications and use them for scheduling decision. Third, we study constraint-aware VM placement in heterogeneous clusters. We present a modelofVMplacementconstraintsandconstraint-awareVMplacementalgorithms. Weanalyze the e???ect of VM placement constraint, and evaluate the performance of algorithms over various settings with simulation and experiments in a small cluster. Finally, we propose interference-awareresource management system for CNN models in GPU cluster. We analyze the e???ect of interference between CNN models. We then propose techniques to mitigate slowdown from interference for target model, and to predict performance of CNN models when they are co-located. We propose heuristic algorithm to schedule CNN models, and evaluate the techniques and algorithm from experiments in GPU cluster.clos
    corecore