6 research outputs found

    What does fault tolerant Deep Learning need from MPI?

    Full text link
    Deep Learning (DL) algorithms have become the de facto Machine Learning (ML) algorithm for large scale data analysis. DL algorithms are computationally expensive - even distributed DL implementations which use MPI require days of training (model learning) time on commonly studied datasets. Long running DL applications become susceptible to faults - requiring development of a fault tolerant system infrastructure, in addition to fault tolerant DL algorithms. This raises an important question: What is needed from MPI for de- signing fault tolerant DL implementations? In this paper, we address this problem for permanent faults. We motivate the need for a fault tolerant MPI specification by an in-depth consideration of recent innovations in DL algorithms and their properties, which drive the need for specific fault tolerance features. We present an in-depth discussion on the suitability of different parallelism types (model, data and hybrid); a need (or lack thereof) for check-pointing of any critical data structures; and most importantly, consideration for several fault tolerance proposals (user-level fault mitigation (ULFM), Reinit) in MPI and their applicability to fault tolerant DL implementations. We leverage a distributed memory implementation of Caffe, currently available under the Machine Learning Toolkit for Extreme Scale (MaTEx). We implement our approaches by ex- tending MaTEx-Caffe for using ULFM-based implementation. Our evaluation using the ImageNet dataset and AlexNet, and GoogLeNet neural network topologies demonstrates the effectiveness of the proposed fault tolerant DL implementation using OpenMPI based ULFM

    Artificial Neural Network Pruning to Extract Knowledge

    Full text link
    Artificial Neural Networks (NN) are widely used for solving complex problems from medical diagnostics to face recognition. Despite notable successes, the main disadvantages of NN are also well known: the risk of overfitting, lack of explainability (inability to extract algorithms from trained NN), and high consumption of computing resources. Determining the appropriate specific NN structure for each problem can help overcome these difficulties: Too poor NN cannot be successfully trained, but too rich NN gives unexplainable results and may have a high chance of overfitting. Reducing precision of NN parameters simplifies the implementation of these NN, saves computing resources, and makes the NN skills more transparent. This paper lists the basic NN simplification problems and controlled pruning procedures to solve these problems. All the described pruning procedures can be implemented in one framework. The developed procedures, in particular, find the optimal structure of NN for each task, measure the influence of each input signal and NN parameter, and provide a detailed verbal description of the algorithms and skills of NN. The described methods are illustrated by a simple example: the generation of explicit algorithms for predicting the results of the US presidential election.Comment: IJCNN 202

    ๋”ฅ๋Ÿฌ๋‹ ๋ฐฉ๋ฒ•๋ก ์„ ์ด์šฉํ•œ ๋†’์€ ์ ์šฉ์„ฑ์„ ๊ฐ€์ง„ ์ˆ˜๊ฒฝ์žฌ๋ฐฐ ํŒŒํ”„๋ฆฌ์นด ๋Œ€์ƒ ์ ˆ์ฐจ ๊ธฐ๋ฐ˜ ๋ชจ๋ธ ๊ฐœ๋ฐœ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๋†์—…์ƒ๋ช…๊ณผํ•™๋Œ€ํ•™ ๋†๋ฆผ์ƒ๋ฌผ์ž์›ํ•™๋ถ€, 2022. 8. ์†์ •์ต.Many agricultural challenges are entangled in a complex interaction between crops and the environment. As a simplifying tool, crop modeling is a process of abstracting and interpreting agricultural phenomena. Understanding based on this interpretation can play a role in supporting academic and social decisions in agriculture. Process-based crop models have solved the challenges for decades to enhance the productivity and quality of crop production; the remaining objectives have led to demand for crop models handling multidirectional analyses with multidimensional information. As a possible milestone to satisfy this goal, deep learning algorithms have been introduced to the complicated tasks in agriculture. However, the algorithms could not replace existing crop models because of the research fragmentation and low accessibility of the crop models. This study established a developmental protocol for a process-based crop model with deep learning methodology. Literature Review introduced deep learning and crop modeling, and it explained the reasons for the necessity of this protocol despite numerous deep learning applications for agriculture. Base studies were conducted with several greenhouse data in Chapters 1 and 2: transfer learning and U-Net structure were utilized to construct an infrastructure for the deep learning application; HyperOpt, a Bayesian optimization method, was tested to calibrate crop models to compare the existing crop models with the developed model. Finally, the process-based crop model with full deep neural networks, DeepCrop, was developed with an attention mechanism and multitask decoders for hydroponic sweet peppers (Capsicum annuum var. annuum) in Chapter 3. The methodology for data integrity showed adequate accuracy, so it was applied to the data in all chapters. HyperOpt was able to calibrate food and feed crop models for sweet peppers. Therefore, the compared models in the final chapter were optimized using HyperOpt. DeepCrop was trained to simulate several growth factors with environment data. The trained DeepCrop was evaluated with unseen data, and it showed the highest modeling efficiency (=0.76) and the lowest normalized root mean squared error (=0.18) than the compared models. With the high adaptability of DeepCrop, it can be used for studies on various scales and purposes. Since all methods adequately solved the given tasks and underlay the DeepCrop development, the established protocol can be a high throughput for enhancing accessibility of crop models, resulting in unifying crop modeling studies.๋†์—… ์‹œ์Šคํ…œ์—์„œ ๋ฐœ์ƒํ•˜๋Š” ๋ฌธ์ œ๋“ค์€ ์ž‘๋ฌผ๊ณผ ํ™˜๊ฒฝ์˜ ์ƒํ˜ธ์ž‘์šฉ ํ•˜์— ๋ณต์žกํ•˜๊ฒŒ ์–ฝํ˜€ ์žˆ๋‹ค. ์ž‘๋ฌผ ๋ชจ๋ธ๋ง์€ ๋Œ€์ƒ์„ ๋‹จ์ˆœํ™”ํ•˜๋Š” ๋ฐฉ๋ฒ•์œผ๋กœ์จ, ๋†์—…์—์„œ ์ผ์–ด๋‚˜๋Š” ํ˜„์ƒ์„ ์ถ”์ƒํ™”ํ•˜๊ณ  ํ•ด์„ํ•˜๋Š” ๊ณผ์ •์ด๋‹ค. ๋ชจ๋ธ๋ง์„ ํ†ตํ•ด ๋Œ€์ƒ์„ ์ดํ•ดํ•˜๋Š” ๊ฒƒ์€ ๋†์—… ๋ถ„์•ผ์˜ ํ•™์ˆ ์  ๋ฐ ์‚ฌํšŒ์  ๊ฒฐ์ •์„ ์ง€์›ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ง€๋‚œ ์ˆ˜๋…„ ๊ฐ„ ์ ˆ์ฐจ ๊ธฐ๋ฐ˜ ์ž‘๋ฌผ ๋ชจ๋ธ์€ ๋†์—…์˜ ๋ฌธ์ œ๋“ค์„ ํ•ด๊ฒฐํ•˜์—ฌ ์ž‘๋ฌผ ์ƒ์‚ฐ์„ฑ ๋ฐ ํ’ˆ์งˆ์„ ์ฆ์ง„์‹œ์ผฐ์œผ๋ฉฐ, ํ˜„์žฌ ์ž‘๋ฌผ ๋ชจ๋ธ๋ง์— ๋‚จ์•„์žˆ๋Š” ๊ณผ์ œ๋“ค์€ ๋‹ค์ฐจ์› ์ •๋ณด๋ฅผ ๋‹ค๋ฐฉํ–ฅ์—์„œ ๋ถ„์„ํ•  ์ˆ˜ ์žˆ๋Š” ์ž‘๋ฌผ ๋ชจ๋ธ์„ ํ•„์š”๋กœ ํ•˜๊ฒŒ ๋˜์—ˆ๋‹ค. ์ด๋ฅผ ๋งŒ์กฑ์‹œํ‚ฌ ์ˆ˜ ์žˆ๋Š” ์ง€์นจ์œผ๋กœ์จ, ๋ณต์žกํ•œ ๋†์—…์  ๊ณผ์ œ๋“ค์„ ๋ชฉํ‘œ๋กœ ๋”ฅ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜์ด ๋„์ž…๋˜์—ˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜, ์ด ์•Œ๊ณ ๋ฆฌ์ฆ˜๋“ค์€ ๋‚ฎ์€ ๋ฐ์ดํ„ฐ ์™„๊ฒฐ์„ฑ ๋ฐ ๋†’์€ ์—ฐ๊ตฌ ๋‹ค์–‘์„ฑ ๋•Œ๋ฌธ์— ๊ธฐ์กด์˜ ์ž‘๋ฌผ ๋ชจ๋ธ๋“ค์„ ๋Œ€์ฒดํ•˜์ง€๋Š” ๋ชปํ–ˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ์—์„œ๋Š” ๋”ฅ๋Ÿฌ๋‹ ๋ฐฉ๋ฒ•๋ก ์„ ์ด์šฉํ•˜์—ฌ ์ ˆ์ฐจ ๊ธฐ๋ฐ˜ ์ž‘๋ฌผ ๋ชจ๋ธ์„ ๊ตฌ์ถ•ํ•˜๋Š” ๊ฐœ๋ฐœ ํ”„๋กœํ† ์ฝœ์„ ํ™•๋ฆฝํ•˜์˜€๋‹ค. Literature Review์—์„œ๋Š” ๋”ฅ๋Ÿฌ๋‹๊ณผ ์ž‘๋ฌผ ๋ชจ๋ธ์— ๋Œ€ํ•ด ์†Œ๊ฐœํ•˜๊ณ , ๋†์—…์œผ๋กœ์˜ ๋”ฅ๋Ÿฌ๋‹ ์ ์šฉ ์—ฐ๊ตฌ๊ฐ€ ๋งŽ์Œ์—๋„ ์ด ํ”„๋กœํ† ์ฝœ์ด ํ•„์š”ํ•œ ์ด์œ ๋ฅผ ์„ค๋ช…ํ•˜์˜€๋‹ค. ์ œ1์žฅ๊ณผ 2์žฅ์—์„œ๋Š” ๊ตญ๋‚ด ์—ฌ๋Ÿฌ ์ง€์—ญ์˜ ๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜์—ฌ ์ „์ด ํ•™์Šต ๋ฐ U-Net ๊ตฌ์กฐ๋ฅผ ํ™œ์šฉํ•˜์—ฌ ๋”ฅ๋Ÿฌ๋‹ ๋ชจ๋ธ ์ ์šฉ์„ ์œ„ํ•œ ๊ธฐ๋ฐ˜์„ ๋งˆ๋ จํ•˜๊ณ , ๋ฒ ์ด์ง€์•ˆ ์ตœ์ ํ™” ๋ฐฉ๋ฒ•์ธ HyperOpt๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ๊ธฐ์กด ๋ชจ๋ธ๊ณผ ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฐ˜ ๋ชจ๋ธ์„ ๋น„๊ตํ•˜๊ธฐ ์œ„ํ•ด ์‹œํ—˜์ ์œผ๋กœ WOFOST ์ž‘๋ฌผ ๋ชจ๋ธ์„ ๋ณด์ •ํ•˜๋Š” ๋“ฑ ๋ชจ๋ธ ๊ฐœ๋ฐœ์„ ์œ„ํ•œ ๊ธฐ๋ฐ˜ ์—ฐ๊ตฌ๋ฅผ ์ˆ˜ํ–‰ํ•˜์˜€๋‹ค. ๋งˆ์ง€๋ง‰์œผ๋กœ, ์ œ3์žฅ์—์„œ๋Š” ์ฃผ์˜ ๋ฉ”์ปค๋‹ˆ์ฆ˜ ๋ฐ ๋‹ค์ค‘ ์ž‘์—… ๋””์ฝ”๋”๋ฅผ ๊ฐ€์ง„ ์™„์ „ ์‹ฌ์ธต ์‹ ๊ฒฝ๋ง ์ ˆ์ฐจ ๊ธฐ๋ฐ˜ ์ž‘๋ฌผ ๋ชจ๋ธ์ธ DeepCrop์„ ์ˆ˜๊ฒฝ์žฌ๋ฐฐ ํŒŒํ”„๋ฆฌ์นด(Capsicum annuum var. annuum) ๋Œ€์ƒ์œผ๋กœ ๊ฐœ๋ฐœํ•˜์˜€๋‹ค. ๋ฐ์ดํ„ฐ ์™„๊ฒฐ์„ฑ์„ ์œ„ํ•œ ๊ธฐ์ˆ ๋“ค์€ ์ ํ•ฉํ•œ ์ •ํ™•๋„๋ฅผ ๋ณด์—ฌ์ฃผ์—ˆ์œผ๋ฉฐ, ์ „์ฒด ์ฑ•ํ„ฐ ๋ฐ์ดํ„ฐ์— ์ ์šฉํ•˜์˜€๋‹ค. HyperOpt๋Š” ์‹๋Ÿ‰ ๋ฐ ์‚ฌ๋ฃŒ ์ž‘๋ฌผ ๋ชจ๋ธ๋“ค์„ ํŒŒํ”„๋ฆฌ์นด ๋Œ€์ƒ์œผ๋กœ ๋ณด์ •ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ๋”ฐ๋ผ์„œ, ์ œ3์žฅ์˜ ๋น„๊ต ๋Œ€์ƒ ๋ชจ๋ธ๋“ค์— ๋Œ€ํ•ด HyperOpt๋ฅผ ์‚ฌ์šฉํ•˜์˜€๋‹ค. DeepCrop์€ ํ™˜๊ฒฝ ๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜๊ณ  ์—ฌ๋Ÿฌ ์ƒ์œก ์ง€ํ‘œ๋ฅผ ์˜ˆ์ธกํ•˜๋„๋ก ํ•™์Šต๋˜์—ˆ๋‹ค. ํ•™์Šต์— ์‚ฌ์šฉํ•˜์ง€ ์•Š์€ ๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜์—ฌ ํ•™์Šต๋œ DeepCrop๋ฅผ ํ‰๊ฐ€ํ•˜์˜€์œผ๋ฉฐ, ์ด ๋•Œ ๋น„๊ต ๋ชจ๋ธ๋“ค ์ค‘ ๊ฐ€์žฅ ๋†’์€ ๋ชจํ˜• ํšจ์œจ(EF=0.76)๊ณผ ๊ฐ€์žฅ ๋‚ฎ์€ ํ‘œ์ค€ํ™” ํ‰๊ท  ์ œ๊ณฑ๊ทผ ์˜ค์ฐจ(NRMSE=0.18)๋ฅผ ๋ณด์—ฌ์ฃผ์—ˆ๋‹ค. DeepCrop์€ ๋†’์€ ์ ์šฉ์„ฑ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ๋‹ค์–‘ํ•œ ๋ฒ”์œ„์™€ ๋ชฉ์ ์„ ๊ฐ€์ง„ ์—ฐ๊ตฌ์— ์‚ฌ์šฉ๋  ์ˆ˜ ์žˆ์„ ๊ฒƒ์ด๋‹ค. ๋ชจ๋“  ๋ฐฉ๋ฒ•๋“ค์ด ์ฃผ์–ด์ง„ ์ž‘์—…์„ ์ ์ ˆํžˆ ํ’€์–ด๋ƒˆ๊ณ  DeepCrop ๊ฐœ๋ฐœ์˜ ๊ทผ๊ฑฐ๊ฐ€ ๋˜์—ˆ์œผ๋ฏ€๋กœ, ๋ณธ ๋…ผ๋ฌธ์—์„œ ํ™•๋ฆฝํ•œ ํ”„๋กœํ† ์ฝœ์€ ์ž‘๋ฌผ ๋ชจ๋ธ์˜ ์ ‘๊ทผ์„ฑ์„ ํ–ฅ์ƒ์‹œํ‚ฌ ์ˆ˜ ์žˆ๋Š” ํš๊ธฐ์ ์ธ ๋ฐฉํ–ฅ์„ ์ œ์‹œํ•˜์˜€๊ณ , ์ž‘๋ฌผ ๋ชจ๋ธ ์—ฐ๊ตฌ์˜ ํ†ตํ•ฉ์— ๊ธฐ์—ฌํ•  ์ˆ˜ ์žˆ์„ ๊ฒƒ์œผ๋กœ ๊ธฐ๋Œ€ํ•œ๋‹ค.LITERATURE REVIEW 1 ABSTRACT 1 BACKGROUND 3 REMARKABLE APPLICABILITY AND ACCESSIBILITY OF DEEP LEARNING 12 DEEP LEARNING APPLICATIONS FOR CROP PRODUCTION 17 THRESHOLDS TO APPLY DEEP LEARNING TO CROP MODELS 18 NECESSITY TO PRIORITIZE DEEP-LEARNING-BASED CROP MODELS 20 REQUIREMENTS OF THE DEEP-LEARNING-BASED CROP MODELS 21 OPENING REMARKS AND THESIS OBJECTIVES 22 LITERATURE CITED 23 Chapter 1 34 Chapter 1-1 35 ABSTRACT 35 INTRODUCTION 37 MATERIALS AND METHODS 40 RESULTS 50 DISCUSSION 59 CONCLUSION 63 LITERATURE CITED 64 Chapter 1-2 71 ABSTRACT 71 INTRODUCTION 73 MATERIALS AND METHODS 75 RESULTS 84 DISCUSSION 92 CONCLUSION 101 LITERATURE CITED 102 Chapter 2 108 ABSTRACT 108 NOMENCLATURE 110 INTRODUCTION 112 MATERIALS AND METHODS 115 RESULTS 124 DISCUSSION 133 CONCLUSION 137 LITERATURE CITED 138 Chapter 3 144 ABSTRACT 144 INTRODUCTION 146 MATERIALS AND METHODS 149 RESULTS 169 DISCUSSION 182 CONCLUSION 187 LITERATURE CITED 188 GENERAL DISCUSSION 196 GENERAL CONCLUSION 201 ABSTRACT IN KOREAN 203 APPENDIX 204๋ฐ•
    corecore