38 research outputs found

    Enhanced CNN for image denoising

    Full text link
    Owing to flexible architectures of deep convolutional neural networks (CNNs), CNNs are successfully used for image denoising. However, they suffer from the following drawbacks: (i) deep network architecture is very difficult to train. (ii) Deeper networks face the challenge of performance saturation. In this study, the authors propose a novel method called enhanced convolutional neural denoising network (ECNDNet). Specifically, they use residual learning and batch normalisation techniques to address the problem of training difficulties and accelerate the convergence of the network. In addition, dilated convolutions are used in the proposed network to enlarge the context information and reduce the computational cost. Extensive experiments demonstrate that the ECNDNet outperforms the state-of-the-art methods for image denoising.Comment: CAAI Transactions on Intelligence Technology[J], 201

    ์‹ฌ์ธต ์‹ ๊ฒฝ๋ง์„ ํ™œ์šฉํ•œ ์ž๋™ ์˜์ƒ ์žก์Œ ์ œ๊ฑฐ ๊ธฐ๋ฒ•

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ (๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› : ์ž์—ฐ๊ณผํ•™๋Œ€ํ•™ ํ˜‘๋™๊ณผ์ • ๊ณ„์‚ฐ๊ณผํ•™์ „๊ณต, 2020. 8. ๊ฐ•๋ช…์ฃผ.Noise removal in digital image data is a fundamental and important task in the field of image processing. The goal of the task is to remove noises from the given degraded images while maintaining essential details such as edges, curves, textures, etc. There have been various attempts on image denoising: mainly model-based methods such as filtering methods, total variation based methods, non-local mean based approaches. Deep learning have been attracting signi๏ฌcant research interest as they have shown better results than the classical methods in almost all fields. Deep learning-based methods use a large amount of data to train a network for its own objective; in the image denoising case, in order to map the corrupted image to a desired clean image. In this thesis we proposed a new network architecture focusing on white Gaussian noise and real noise cancellation. Our model is a deep and wide network designed by constructing a basic block consisting of a mixture of various types of dilated convolutions and repeatedly stacking them. We did not use a batch normal layer to maintain the original own color information of each input data. Also skip connection was utilized so as not to lose the existing information. Through several experiments and comparisons, it was proved that the proposed network has better performance compared to the traditional and latest methods in image denoising.๋””์ง€ํ„ธ ์˜์ƒ ๋ฐ์ดํ„ฐ ๋‚ด์˜ ์žก์Œ ์ œ๊ฑฐ ๋ฐ ๊ฐ์†Œ๋Š” ์—ดํ™”๋œ ์˜์ƒ์˜ ๋…ธ์ด์ฆˆ๋ฅผ ์ œ๊ฑฐํ•˜๋ฉด์„œ ๋ชจ์„œ๋ฆฌ, ๊ณก์„ , ์งˆ๊ฐ ๋“ฑ๊ณผ ๊ฐ™์€ ํ•„์ˆ˜ ์„ธ๋ถ€ ์ •๋ณด๋ฅผ ์œ ์ง€ํ•˜๋Š” ๊ฒƒ์ด ๋ชฉ์ ์ธ ์˜์ƒ ์ฒ˜๋ฆฌ ๋ถ„์•ผ์˜ ๊ธฐ๋ณธ์ ์ด๊ณ  ํ•„์ˆ˜์ ์ธ ์ž‘์—…์ด๋‹ค. ๋”ฅ๋Ÿฌ๋‹ ๊ธฐ๋ฐ˜์˜ ์˜์ƒ ์žก์Œ ์ œ๊ฑฐ ๋ฐฉ๋ฒ•๋“ค์€ ์—ดํ™”๋œ ์˜์ƒ์„ ์›ํ•˜๋Š” ํ’ˆ์งˆ์˜ ์˜์ƒ์œผ๋กœ ๋งคํ•‘ํ•˜๋„๋ก ๋Œ€์šฉ๋Ÿ‰์˜ ๋ฐ์ดํ„ฐ๋ฅผ ์ด์šฉํ•˜์—ฌ ๋„คํŠธ์›Œํฌ๋ฅผ ์ง€๋„ํ•™์Šตํ•˜๋ฉฐ ๊ณ ์ „์ ์ธ ๋ฐฉ๋ฒ•๋“ค๋ณด๋‹ค ๋›ฐ์–ด๋‚œ ๊ฒฐ๊ณผ๋ฅผ ๋ณด์—ฌ์ฃผ๊ณ  ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์ด๋ฏธ์ง€ ๋””๋…ธ์ด์ง•์— ๋Œ€ํ•œ ์—ฌ๋Ÿฌ ๋ฐฉ๋ฒ•๋“ค์„ ์กฐ์‚ฌํ–ˆ์„ ๋ฟ๋งŒ ์•„๋‹ˆ๋ผ, ํŠนํžˆ ๋ฐฑ์ƒ‰ ๊ฐ€์šฐ์‹œ์•ˆ ์žก์Œ๊ณผ ์‹ค์ œ ์žก์Œ ์ œ๊ฑฐ ๋ฌธ์ œ์— ์ง‘์ค‘ํ•˜๋ฉด์„œ ๋„คํŠธ์›Œํฌ ์•„ํ‚คํ…์ฒ˜๋ฅผ ์„ค๊ณ„ํ•˜๊ณ  ์‹คํ—˜ํ•˜์˜€๋‹ค. ์—ฌ๋Ÿฌ ํ˜•ํƒœ์˜ ๋”œ๋ ˆ์ดํ‹ฐ๋“œ ์ฝ˜๋ณผ๋ฃจ์…˜๋“ค์„ ํ˜ผํ•ฉํ•˜์—ฌ ๊ธฐ๋ณธ ๋ธ”๋ก์„ ๊ตฌ์„ฑํ•˜๊ณ  ์ด๋ฅผ ๋ฐ˜๋ณตํ•˜์—ฌ ์Œ“์•„์„œ ์„ค๊ณ„ํ•œ ๋„คํŠธ์›Œํฌ๋ฅผ ์ œ์•ˆํ•˜์˜€๊ณ , ๊ฐ๊ฐ ๋ณธ์—ฐ์˜ ์ƒ‰์ƒ์„ ์œ ์ง€ํ•  ์ˆ˜ ์žˆ๋„๋ก ์—ฌ๋Ÿฌ ์ž…๋ ฅ ์˜์ƒ์„ ํ•˜๋‚˜๋กœ ๋ฌถ์–ด ๊ตฌ์„ฑํ•˜๋Š” ๋ฐฐ์น˜๋ฅผ ํ‰์ค€ํ™”ํ•˜๋Š” ๋ฐฐ์น˜๋…ธ๋ฉ€ ๋ ˆ์ด์–ด๋Š” ์‚ฌ์šฉํ•˜์ง€ ์•Š์•˜๋‹ค. ๊ทธ๋ฆฌ๊ณ  ๋ธ”๋ก์ด ์—ฌ๋Ÿฌ ์ธต ์ง„ํ–‰๋˜๋Š” ๋™์•ˆ ๊ธฐ์กด์˜ ์ •๋ณด๋ฅผ ์†์‹คํ•˜์ง€ ์•Š๋„๋ก ์Šคํ‚ต ์ปค๋„ฅ์…˜์„ ์‚ฌ์šฉํ•˜์˜€๋‹ค. ์—ฌ๋Ÿฌ ์‹คํ—˜๊ณผ ๊ธฐ์กด์— ์ œ์•ˆ๋œ ๋ฐฉ๋ฒ•๊ณผ ์ตœ์‹  ๋ฒค์น˜ ๋งˆํฌ์™€์˜ ๋น„๊ต๋ฅผ ํ†ตํ•˜์—ฌ ์ œ์•ˆํ•œ ๋„คํŠธ์›Œํฌ๊ฐ€ ๋…ธ์ด์ฆˆ ๊ฐ์†Œ ๋ฐ ์ œ๊ฑฐ ์ž‘์—…์—์„œ ๊ธฐ์กด์˜ ๋ฐฉ๋ฒ•๋“ค๊ณผ ๋น„๊ตํ•˜์—ฌ ์šฐ์ˆ˜ํ•œ ์„ฑ๋Šฅ์„ ๊ฐ€์ง€๊ณ  ์žˆ์Œ์„ ์ž…์ฆํ•˜์˜€๋‹ค. ํ•˜์ง€๋งŒ ์ œ์•ˆํ•œ ์•„ํ‚คํ…์ฒ˜๋ฐฉ๋ฒ•์˜ ํ•œ๊ณ„์ ๋„ ๋ช‡ ๊ฐ€์ง€ ์กด์žฌํ•œ๋‹ค. ์ œ์•ˆํ•œ ๋„คํŠธ์›Œํฌ๋Š” ๋‹ค์šด์ƒ˜ํ”Œ๋ง์„ ์‚ฌ์šฉํ•˜์ง€ ์•Š์Œ์œผ๋กœ์จ ์ •๋ณด ์†์‹ค์„ ์ตœ์†Œํ™”ํ•˜์˜€์ง€๋งŒ ์ตœ์‹  ๋ฒค์น˜๋งˆํฌ์— ๋น„ํ•˜์—ฌ ๋” ๋งŽ์€ ์ถ”๋ก  ์‹œ๊ฐ„์ด ํ•„์š”ํ•˜์—ฌ ์‹ค์‹œ๊ฐ„ ์ž‘์—…์—๋Š” ์ ์šฉํ•˜๊ธฐ๊ฐ€ ์‰ฝ์ง€ ์•Š๋‹ค. ์‹ค์ œ ์˜์ƒ์—๋Š” ๋‹จ์ˆœํ•œ ์žก์Œ๋ณด๋‹ค๋Š” ์˜์ƒ ํš๋“, ์ €์žฅ ๋“ฑ๊ณผ ๊ฐ™์€ ํ”„๋กœ์„ธ์Šค๋ฅผ ๊ฑฐ์น˜๋ฉด์„œ ์—ฌ๋Ÿฌ ์š”์ธ๋“ค๋กœ ์ธํ•œ ๋‹ค์–‘ํ•œ ์žก์Œ, ๋ธ”๋Ÿฌ์™€ ๊ฐ™์€ ์—ดํ™”๊ฐ€ ํ˜ผ์žฌ ๋˜์–ด ์žˆ๋‹ค. ์‹ค์ œ ์žก์Œ์— ๋Œ€ํ•œ ๋‹ค์–‘ํ•œ ๊ฐ๋„์—์„œ์˜ ๋ถ„์„๊ณผ ์—ฌ๋Ÿฌ ๋ชจ๋ธ๋ง ์‹คํ—˜, ๊ทธ๋ฆฌ๊ณ  ์˜์ƒ ์žก์Œ ๋ฐ ๋ธ”๋Ÿฌ, ์••์ถ•๊ณผ ๊ฐ™์€ ๋ณตํ•ฉ ๋ชจ๋ธ๋ง์ด ํ•„์š”ํ•˜๋‹ค. ํ–ฅํ›„์—๋Š” ์ด๋Ÿฌํ•œ ์ ๋“ค์„ ๋ณด์™„ํ•จ์œผ๋กœ์จ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚ค๊ณ  ๋„คํŠธ์›Œํฌ์˜ ์กฐ์ •์„ ํ†ตํ•ด ์‹ค์‹œ๊ฐ„์œผ๋กœ ์ ์šฉ๋  ์ˆ˜ ์žˆ์Œ์„ ๊ธฐ๋Œ€ํ•œ๋‹ค.1 Introduction 1 2 Review on Image Denoising Methods 4 2.1 Image Noise Models 4 2.2 Traditional Denoising Methods 8 2.2.1 TV-based regularization 9 2.2.2 Non-local regularization 9 2.2.3 Sparse representation 10 2.2.4 Low-rank minimization 10 2.3 CNN-based Denoising Methods 11 2.3.1 DnCNN 11 2.3.2 FFDNet 12 2.3.3 WDnCNN 12 2.3.4 DHDN 13 3 Proposed models 15 3.1 Related Works 15 3.1.1 Residual learning 15 3.1.2 Dilated convolution 16 3.2 Proposed Network Architecture 17 4 Experiments 21 4.1 Training Details 21 4.2 Synthetic Noise Reduction 23 4.2.1 Set12 denoising 24 4.2.2 Kodak24 and BSD68 denoising 30 4.3 Real Noise Reduction 34 4.3.1 DnD test results 35 4.3.2 NTIRE 2020 real image denoising challenge 42 5 Conclusion and Future Works 46 Abstract (in Korean) 54Docto

    Fully Trainable and Interpretable Non-Local Sparse Models for Image Restoration

    Get PDF
    Non-local self-similarity and sparsity principles have proven to be powerful priors for natural image modeling. We propose a novel differentiable relaxation of joint sparsity that exploits both principles and leads to a general framework for image restoration which is (1) trainable end to end, (2) fully interpretable, and (3) much more compact than competing deep learning architectures. We apply this approach to denoising, jpeg deblocking, and demosaicking, and show that, with as few as 100K parameters, its performance on several standard benchmarks is on par or better than state-of-the-art methods that may have an order of magnitude or more parameters.Comment: ECCV 202
    corecore