16,527 research outputs found

    ์‹ฌ์ธต ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰ ๊ธฐ๋ฒ•์„ ์‚ฌ์šฉํ•œ ์ด๋ฏธ์ง€ ๋ณต์›

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ „๊ธฐยท์ •๋ณด๊ณตํ•™๋ถ€, 2021.8. ์•ˆ์ค€์˜.Image restoration is an important technology which can be used as a pre-processing step to increase the performances of various vision tasks. Image super-resolution is one of the important task in image restoration which restores a high-resolution (HR) image from low-resolution (LR) observation. The recent progress of deep convolutional neural networks has enabled great success in single image super-resolution (SISR). its performance is also being increased by deepening the networks and developing more sophisticated network structures. However, finding an optimal structure for the given problem is a difficult task, even for human experts. For this reason, neural architecture search (NAS) methods have been introduced, which automate the procedure of constructing the structures. In this dissertation, I propose a new single image super-resolution framework by using neural architecture search (NAS) method. As the performance improves, the network becomes more complex and deeper, so I apply NAS algorithm to find the optimal network while reducing the effort in network design. In detail, the proposed scheme is summarized to three topics: image super-resolution using efficient neural architecture search, multi-branch neural architecture search for lightweight image super-resolution, and neural architecture search for image super-resolution using meta-transfer learning. At first, I expand the NAS to the super-resolution domain and find a lightweight densely connected network named DeCoNASNet. I use a hierarchical search strategy to find the best connection with local and global features. In this process, I define a complexity-based-penalty and add it to the reward term of REINFORCE algorithm. Experiments show that my DeCoNASNet outperforms the state-of-the-art lightweight super-resolution networks designed by handcraft methods and existing NAS-based design. I propose a new search space design with multi-branch structure to enlarge the search space for capturing multi-scale features, resulting in better reconstruction on grainy areas. I also adopt parameter sharing scheme in multi-branch network to share their information and reduce the whole network parameter. Experiments show that the proposed method finds an optimal SISR network about twenty times faster than the existing methods, while showing comparable performance in terms of PSNR vs. parameters. Comparison of visual quality validates that the proposed SISR network reconstructs texture areas better than the previous methods because of the enlarged search space to find multi-scale features. Lastly, I apply meta-transfer learning to the NAS procedure for image super-resolution. I train the controller and child network with the meta-learning scheme, which enables the controllers to find promising network for several scale simultaneously. Furthermore, meta-trained child network is reused as the pre-trained parameters for final evaluation phase to improve the final image super-resolution results even better and search-evaluation gap problem is efficiently reduced.์ด๋ฏธ์ง€ ๋ณต์›์€ ๋‹ค์–‘ํ•œ ์˜์ƒ์ฒ˜๋ฆฌ ๋ฌธ์ œ์˜ ์„ฑ๋Šฅ์„ ๋†’์ด๊ธฐ ์œ„ํ•œ ์ „ ์ฒ˜๋ฆฌ ๋‹จ๊ณ„๋กœ ์‚ฌ์šฉํ•  ์ˆ˜ ์žˆ๋Š” ์ค‘์š”ํ•œ ๊ธฐ์ˆ ์ด๋‹ค. ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™”๋Š” ์ด๋ฏธ์ง€ ๋ณต์›๋ฐฉ๋ฒ• ์ค‘ ์ค‘์š”ํ•œ ๋ฌธ์ œ์˜ ํ•˜๋‚˜๋กœ์จ ์ €ํ•ด์ƒ๋„์˜ ์ด๋ฏธ์ง€๋ฅผ ๊ณ ํ•ด์ƒ๋„์˜ ์ด๋ฏธ์ง€๋กœ ๋ณต์›ํ•˜๋Š” ๋ฐฉ๋ฒ•์ด๋‹ค. ์ตœ๊ทผ์—๋Š” ์ปจ๋ฒŒ๋ฃจ์…˜ ์‹ ๊ฒฝ๋ง (CNN)์„ ์‚ฌ์šฉํ•˜๋Š” ๋”ฅ ๋Ÿฌ๋‹(deep learning) ๊ธฐ๋ฐ˜์˜ ๋ฐฉ๋ฒ•๋“ค์ด ๋‹จ์ผ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” (SISR) ๋ฌธ์ œ๋ฅผ ํ‘ธ๋Š”๋ฐ ๋งŽ์ด ์‚ฌ์šฉ๋˜๊ณ  ์žˆ๋‹ค. ์ผ๋ฐ˜์ ์œผ๋กœ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” ์„ฑ๋Šฅ์€ CNN์„ ๊นŠ๊ฒŒ ์Œ“๊ฑฐ๋‚˜ ๋ณต์žกํ•œ ๊ตฌ์กฐ๋ฅผ ์„ค๊ณ„ํ•จ์œผ๋กœ์จ ํ–ฅ์ƒ์‹œํ‚ฌ ์ˆ˜ ์žˆ๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ์ฃผ์–ด์ง„ ๋ฌธ์ œ์— ๋Œ€ํ•œ ์ตœ์ ์˜ ๊ตฌ์กฐ๋ฅผ ์ฐพ๋Š” ๊ฒƒ์€ ํ•ด๋‹น ๋ถ„์•ผ์˜ ์ „๋ฌธ๊ฐ€๋ผ ํ•  ์ง€๋ผ๋„ ์–ด๋ ต๊ณ  ์‹œ๊ฐ„์ด ์˜ค๋ž˜ ๊ฑธ๋ฆฌ๋Š” ์ž‘์—…์ด๋‹ค. ์ด๋Ÿฌํ•œ ์ด์œ ๋กœ ์‹ ๊ฒฝ๋ง ๊ตฌ์ถ• ์ ˆ์ฐจ๋ฅผ ์ž๋™ํ™”ํ•˜๋Š” ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ ๊ฒ€์ƒ‰ (NAS) ๋ฐฉ๋ฒ•์ด ๋„์ž…๋˜์—ˆ๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ๋Š” ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ ๊ฒ€์ƒ‰ (NAS) ๋ฐฉ๋ฒ•์„ ์‚ฌ์šฉํ•˜์—ฌ ์ƒˆ๋กœ์šด ๋‹จ์ผ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ ์ œ์•ˆํ•œ ๋ฐฉ๋ฒ•์€ ํฌ๊ฒŒ ์„ธ ๊ฐ€์ง€๋กœ ์š”์•ฝ ํ•  ์ˆ˜ ์žˆ๋‹ค. ์ด๋Š” ํšจ์œจ์ ์ธ ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰๊ธฐ๋ฒ•(ENAS)์„ ์ด์šฉํ•œ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™”, ๋ณ‘๋ ฌ ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰ ๊ธฐ๋ฒ•์„ ์ด์šฉํ•œ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™”, ๋ฉ”ํƒ€ ์ „์†ก ํ•™์Šต์„ ์ด์šฉํ•˜๋Š” ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰๊ธฐ๋ฒ•์„ ํ†ตํ•œ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” ์ด๋‹ค. ์šฐ์„ , ์šฐ๋ฆฌ๋Š” ์ฃผ๋กœ ์˜์ƒ ๋ถ„๋ฅ˜์— ์“ฐ์ด๋˜ ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰ ๊ธฐ๋ฒ•์„ ์˜์ƒ ๊ณ ํ•ด์ƒ๋„ํ™”์— ์ ์šฉํ•˜์˜€์œผ๋ฉฐ, DeCoNASNet์ด๋ผ ๋ช…๋ช…๋œ ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ๋ฅผ ์„ค๊ณ„ํ•˜์˜€๋‹ค. ๋˜ํ•œ ๊ณ„์ธต์  ๊ฒ€์ƒ‰ ์ „๋žต์„ ์‚ฌ์šฉํ•˜์—ฌ ์ง€์—ญ/์ „์—ญ ํ”ผ์ณ(feature) ํ•ฉ๋ณ‘์„ ์œ„ํ•œ ์ตœ์ƒ์˜ ์—ฐ๊ฒฐ ๋ฐฉ๋ฒ•์„ ๊ฒ€์ƒ‰ํ•˜์˜€๋‹ค. ์ด ๊ณผ์ •์—์„œ ํ•„์š” ๋ณ€์ˆ˜๊ฐ€ ์ ์œผ๋ฉด์„œ ์ข‹์€ ์„ฑ๋Šฅ์„ ๋‚ผ ์ˆ˜ ์žˆ๋„๋ก ๋ณต์žก์„ฑ ๊ธฐ๋ฐ˜ ํŽ˜๋„ํ‹ฐ (complexity-based penalty) ๋ฅผ ์ •์˜ํ•˜๊ณ  ์ด๋ฅผ REINFORCE ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ๋ณด์ƒ ์‹ ํ˜ธ์— ์ถ”๊ฐ€ํ•˜์˜€๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ DeCoNASNet์€ ๊ธฐ์กด์˜ ์‚ฌ๋žŒ์ด ์ง์ ‘ ์„ค๊ณ„ํ•œ ์‹ ๊ฒฝ๋ง๊ณผ ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰ ๊ธฐ๋ฒ•์„ ๊ธฐ๋ฐ˜์œผ๋กœ ์„ค๊ณ„๋œ ์ตœ๊ทผ์˜ ๊ณ ํ•ด์ƒ๋„ํ™” ๊ตฌ์กฐ์˜ ์„ฑ๋Šฅ์„ ๋Šฅ๊ฐ€ํ•˜๋Š” ๊ฒƒ์„ ํ™•์ธ ํ•  ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์šฐ๋ฆฌ๋Š” ๋˜ํ•œ ์—ฌ๋Ÿฌ ํฌ๊ธฐ์˜ ํ”ผ์ณ(feature)๋ฅผ ํ•™์Šตํ•˜๊ธฐ ์œ„ํ•ด ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰ ๊ธฐ๋ฒ•์˜ ๊ฒ€์ƒ‰ ๊ณต๊ฐ„์„ ํ™•๋Œ€ํ•˜์—ฌ ๋ณ‘๋ ฌ ์‹ ๊ฒฝ๋ง์„ ์„ค๊ณ„ํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ด ๋•Œ, ๋ณ‘๋ ฌ์‹ ๊ฒฝ๋ง์˜ ๊ฐ ์œ„์น˜์—์„œ ๋งค๊ฐœ ๋ณ€์ˆ˜๋ฅผ ๊ณต์œ ํ•  ์ˆ˜ ์žˆ๋„๋ก ํ•˜์—ฌ ๋ณ‘๋ ฌ์‹ ๊ฒฝ๋ง์˜ ๊ฐ ๊ตฌ์กฐ๋ผ๋ฆฌ ์ •๋ณด๋ฅผ ๊ณต์œ ํ•˜๊ณ  ์ „์ฒด ๊ตฌ์กฐ๋ฅผ ์„ค๊ณ„ํ•˜๋Š”๋ฐ ํ•„์š”ํ•œ ๋งค๊ฐœ ๋ณ€์ˆ˜๋ฅผ ์ค„์ด๋„๋ก ํ•˜์˜€๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ ์ œ์•ˆ๋œ ๋ฐฉ๋ฒ•์„ ํ†ตํ•ด ๋งค๊ฐœ ๋ณ€์ˆ˜ ํฌ๊ธฐ ๋Œ€๋น„ ์„ฑ๋Šฅ์ด ์ข‹์€ ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ๋ฅผ ์ฐพ์„ ์ˆ˜ ์žˆ์—ˆ๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ๋ฅผ ํ†ตํ•ด ํ™•์žฅ๋œ ๊ฒ€์ƒ‰ ๊ณต๊ฐ„์—์„œ ์—ฌ๋Ÿฌ ํฌ๊ธฐ์˜ ํ”ผ์ณ (feature)๋ฅผ ํ•™์Šตํ•˜์˜€๊ธฐ ๋•Œ๋ฌธ์— ์ด์ „ ๋ฐฉ๋ฒ•๋ณด๋‹ค ๋ณต์žกํ•œ ์˜์—ญ์„ ๋” ์ž˜ ๋ณต์›ํ•˜๋Š” ๊ฒƒ์„ ํ™•์ธํ•˜์˜€๋‹ค. ๋งˆ์ง€๋ง‰์œผ๋กœ ๋ฉ”ํƒ€ ์ „์†ก ํ•™์Šต(meta-transfer learning)์„ ์‹ ๊ฒฝ๋ง ๊ฒ€์ƒ‰์— ์ ์šฉํ•˜์—ฌ ๋‹ค์–‘ํ•œ ํฌ๊ธฐ์˜ ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜๋Š” ๋ฐฉ๋ฒ•์„ ์ œ์•ˆํ•˜์˜€๋‹ค. ์ด ๋…ผ๋ฌธ์—์„œ๋Š” ๋ฉ”ํƒ€ ์ „์†ก ํ•™์Šต ๋ฐฉ๋ฒ•์„ ํ†ตํ•ด ์ œ์–ด๊ธฐ๊ฐ€ ์—ฌ๋Ÿฌ ํฌ๊ธฐ์˜ ์ข‹์€ ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ๋ฅผ ๋™์‹œ์— ์ฐพ์„ ์ˆ˜ ์žˆ๋„๋ก ์„ค๊ณ„ํ•˜์˜€๋‹ค. ๋˜ํ•œ ๋ฉ”ํƒ€ ํ›ˆ๋ จ๋œ ์‹ ๊ฒฝ๋ง ๊ตฌ์กฐ๋Š” ์ตœ์ข… ์„ฑ๋Šฅ ํ‰๊ฐ€ ์‹œ ํ•™์Šต์˜ ์‹œ์ž‘์ ์œผ๋กœ ์žฌ์‚ฌ์šฉ ๋˜์–ด ์ตœ์ข… ์ด๋ฏธ์ง€ ๊ณ ํ•ด์ƒ๋„ํ™” ์„ฑ๋Šฅ์„ ๋”์šฑ ํ–ฅ์ƒ์‹œํ‚ฌ ์ˆ˜ ์žˆ์—ˆ์œผ๋ฉฐ, ํšจ๊ณผ์ ์œผ๋กœ ๊ฒ€์ƒ‰-ํ‰๊ฐ€ ๊ดด๋ฆฌ ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜์˜€๋‹ค.1 INTRODUCTION 1 1.1 contribution 3 1.2 contents 4 2 Neural Architecture Search for Image Super-Resolution Using Densely Constructed Search Space: DeCoNAS 5 2.1 Introduction 5 2.2 Proposed Method 9 2.2.1 Overall structure of DeCoNASNet 9 2.2.2 Constructing the DNB 11 2.2.3 Constructing controller for the DeCoNASNet 13 2.2.4 Training DeCoNAS and complexity-based penalty 13 2.3 Experimental results 15 2.3.1 Settings 15 2.3.2 Results 16 2.3.3 Ablation study 21 2.4 Summary 22 3 Multi-Branch Neural Architecture Search for Lightweight Image Super-resolution 23 3.1 Introduction 23 3.2 Related Work 26 3.2.1 Single image super-resolution 26 3.2.2 Neural architecture search 27 3.2.3 Image super-resolution with neural architecture search 29 3.3 Method 32 3.3.1 Overview of the Proposed MBNAS 32 3.3.2 Controller and complexity-based penalty 33 3.3.3 MBNASNet 35 3.3.4 Multi-scale block with partially shared Nodes 37 3.3.5 MBNAS 38 3.4 datasets and experiments 39 3.4.1 Settings 39 3.4.2 Experiments on single image super-resolution (SISR) 41 3.5 Discussion 48 3.5.1 Effect of the complexity-based penalty to the performance of controller 49 3.5.2 Effect of multi-branch structure and partial parameter sharing scheme 50 3.5.3 Effect of gradient flow control weights and complexity-based penalty coefficient 51 3.6 Summary 52 4 Meta-transfer learning for simultaneous search of various scale image super-resolution 54 4.1 Introduction 54 4.2 Related Work 56 4.2.1 Single image super-resolution 56 4.2.2 Neural architecture search 57 4.2.3 Image super-resolution with neural architecture search 58 4.2.4 Meta-learning 59 4.3 Method 59 4.3.1 Meta-learning 60 4.3.2 Meta-transfer learning 62 4.3.3 Transfer-learning 63 4.4 datasets and experiments 63 4.4.1 Settings 63 4.4.2 Experiments on single image super-resolution(SISR) 64 4.5 Summary 66 5 Conclusion 69 Abstract (In Korean) 80๋ฐ•

    Deep Learning for Single Image Super-Resolution: A Brief Review

    Get PDF
    Single image super-resolution (SISR) is a notoriously challenging ill-posed problem, which aims to obtain a high-resolution (HR) output from one of its low-resolution (LR) versions. To solve the SISR problem, recently powerful deep learning algorithms have been employed and achieved the state-of-the-art performance. In this survey, we review representative deep learning-based SISR methods, and group them into two categories according to their major contributions to two essential aspects of SISR: the exploration of efficient neural network architectures for SISR, and the development of effective optimization objectives for deep SISR learning. For each category, a baseline is firstly established and several critical limitations of the baseline are summarized. Then representative works on overcoming these limitations are presented based on their original contents as well as our critical understandings and analyses, and relevant comparisons are conducted from a variety of perspectives. Finally we conclude this review with some vital current challenges and future trends in SISR leveraging deep learning algorithms.Comment: Accepted by IEEE Transactions on Multimedia (TMM

    Hypernetwork functional image representation

    Full text link
    Motivated by the human way of memorizing images we introduce their functional representation, where an image is represented by a neural network. For this purpose, we construct a hypernetwork which takes an image and returns weights to the target network, which maps point from the plane (representing positions of the pixel) into its corresponding color in the image. Since the obtained representation is continuous, one can easily inspect the image at various resolutions and perform on it arbitrary continuous operations. Moreover, by inspecting interpolations we show that such representation has some properties characteristic to generative models. To evaluate the proposed mechanism experimentally, we apply it to image super-resolution problem. Despite using a single model for various scaling factors, we obtained results comparable to existing super-resolution methods
    • โ€ฆ
    corecore