12 research outputs found

    Generative Adversarial Networks for Visible to Infrared Video Conversion

    Get PDF
    Deep learning models are data driven. For example, the most popular convolutional neural network (CNN) model used for image classification or object detection requires large labeled databases for training to achieve competitive performances. This requirement is not difficult to be satisfied in the visible domain since there are lots of labeled video and image databases available nowadays. However, given the less popularity of infrared (IR) camera, the availability of labeled infrared videos or image databases is limited. Therefore, training deep learning models in infrared domain is still challenging. In this chapter, we applied the pix2pix generative adversarial network (Pix2Pix GAN) and cycle-consistent GAN (Cycle GAN) models to convert visible videos to infrared videos. The Pix2Pix GAN model requires visible-infrared image pairs for training while the Cycle GAN relaxes this constraint and requires only unpaired images from both domains. We applied the two models to an open-source database where visible and infrared videos provided by the signal multimedia and telecommunications laboratory at the Federal University of Rio de Janeiro. We evaluated conversion results by performance metrics including Inception Score (IS), Frechet Inception Distance (FID) and Kernel Inception Distance (KID). Our experiments suggest that cycle-consistent GAN is more effective than pix2pix GAN for generating IR images from optical images

    심측 신경망을 ν™œμš©ν•œ μžλ™ μ˜μƒ 작음 제거 기법

    Get PDF
    ν•™μœ„λ…Όλ¬Έ (박사) -- μ„œμšΈλŒ€ν•™κ΅ λŒ€ν•™μ› : μžμ—°κ³Όν•™λŒ€ν•™ ν˜‘λ™κ³Όμ • 계산과학전곡, 2020. 8. κ°•λͺ…μ£Ό.Noise removal in digital image data is a fundamental and important task in the field of image processing. The goal of the task is to remove noises from the given degraded images while maintaining essential details such as edges, curves, textures, etc. There have been various attempts on image denoising: mainly model-based methods such as filtering methods, total variation based methods, non-local mean based approaches. Deep learning have been attracting significant research interest as they have shown better results than the classical methods in almost all fields. Deep learning-based methods use a large amount of data to train a network for its own objective; in the image denoising case, in order to map the corrupted image to a desired clean image. In this thesis we proposed a new network architecture focusing on white Gaussian noise and real noise cancellation. Our model is a deep and wide network designed by constructing a basic block consisting of a mixture of various types of dilated convolutions and repeatedly stacking them. We did not use a batch normal layer to maintain the original own color information of each input data. Also skip connection was utilized so as not to lose the existing information. Through several experiments and comparisons, it was proved that the proposed network has better performance compared to the traditional and latest methods in image denoising.디지털 μ˜μƒ 데이터 λ‚΄μ˜ 작음 제거 및 κ°μ†ŒλŠ” μ—΄ν™”λœ μ˜μƒμ˜ λ…Έμ΄μ¦ˆλ₯Ό μ œκ±°ν•˜λ©΄μ„œ λͺ¨μ„œλ¦¬, 곑선, 질감 λ“±κ³Ό 같은 ν•„μˆ˜ μ„ΈλΆ€ 정보λ₯Ό μœ μ§€ν•˜λŠ” 것이 λͺ©μ μΈ μ˜μƒ 처리 λΆ„μ•Όμ˜ 기본적이고 ν•„μˆ˜μ μΈ μž‘μ—…μ΄λ‹€. λ”₯λŸ¬λ‹ 기반의 μ˜μƒ 작음 제거 방법듀은 μ—΄ν™”λœ μ˜μƒμ„ μ›ν•˜λŠ” ν’ˆμ§ˆμ˜ μ˜μƒμœΌλ‘œ λ§€ν•‘ν•˜λ„λ‘ λŒ€μš©λŸ‰μ˜ 데이터λ₯Ό μ΄μš©ν•˜μ—¬ λ„€νŠΈμ›Œν¬λ₯Ό μ§€λ„ν•™μŠ΅ν•˜λ©° 고전적인 방법듀보닀 λ›°μ–΄λ‚œ κ²°κ³Όλ₯Ό 보여주고 μžˆλ‹€. λ³Έ λ…Όλ¬Έμ—μ„œλŠ” 이미지 디노이징에 λŒ€ν•œ μ—¬λŸ¬ 방법듀을 μ‘°μ‚¬ν–ˆμ„ 뿐만 μ•„λ‹ˆλΌ, 특히 백색 κ°€μš°μ‹œμ•ˆ 작음과 μ‹€μ œ 작음 제거 λ¬Έμ œμ— μ§‘μ€‘ν•˜λ©΄μ„œ λ„€νŠΈμ›Œν¬ μ•„ν‚€ν…μ²˜λ₯Ό μ„€κ³„ν•˜κ³  μ‹€ν—˜ν•˜μ˜€λ‹€. μ—¬λŸ¬ ν˜•νƒœμ˜ λ”œλ ˆμ΄ν‹°λ“œ μ½˜λ³Όλ£¨μ…˜λ“€μ„ ν˜Όν•©ν•˜μ—¬ κΈ°λ³Έ 블둝을 κ΅¬μ„±ν•˜κ³  이λ₯Ό λ°˜λ³΅ν•˜μ—¬ μŒ“μ•„μ„œ μ„€κ³„ν•œ λ„€νŠΈμ›Œν¬λ₯Ό μ œμ•ˆν•˜μ˜€κ³ , 각각 λ³Έμ—°μ˜ 색상을 μœ μ§€ν•  수 μžˆλ„λ‘ μ—¬λŸ¬ μž…λ ₯ μ˜μƒμ„ ν•˜λ‚˜λ‘œ λ¬Άμ–΄ κ΅¬μ„±ν•˜λŠ” 배치λ₯Ό ν‰μ€€ν™”ν•˜λŠ” λ°°μΉ˜λ…Έλ©€ λ ˆμ΄μ–΄λŠ” μ‚¬μš©ν•˜μ§€ μ•Šμ•˜λ‹€. 그리고 블둝이 μ—¬λŸ¬ μΈ΅ μ§„ν–‰λ˜λŠ” λ™μ•ˆ 기쑴의 정보λ₯Ό μ†μ‹€ν•˜μ§€ μ•Šλ„λ‘ μŠ€ν‚΅ 컀λ„₯μ…˜μ„ μ‚¬μš©ν•˜μ˜€λ‹€. μ—¬λŸ¬ μ‹€ν—˜κ³Ό 기쑴에 μ œμ•ˆλœ 방법과 μ΅œμ‹  벀치 λ§ˆν¬μ™€μ˜ 비ꡐλ₯Ό ν†΅ν•˜μ—¬ μ œμ•ˆν•œ λ„€νŠΈμ›Œν¬κ°€ λ…Έμ΄μ¦ˆ κ°μ†Œ 및 제거 μž‘μ—…μ—μ„œ 기쑴의 방법듀과 λΉ„κ΅ν•˜μ—¬ μš°μˆ˜ν•œ μ„±λŠ₯을 가지고 μžˆμŒμ„ μž…μ¦ν•˜μ˜€λ‹€. ν•˜μ§€λ§Œ μ œμ•ˆν•œ μ•„ν‚€ν…μ²˜λ°©λ²•μ˜ ν•œκ³„μ λ„ λͺ‡ 가지 μ‘΄μž¬ν•œλ‹€. μ œμ•ˆν•œ λ„€νŠΈμ›Œν¬λŠ” λ‹€μš΄μƒ˜ν”Œλ§μ„ μ‚¬μš©ν•˜μ§€ μ•ŠμŒμœΌλ‘œμ¨ 정보 손싀을 μ΅œμ†Œν™”ν•˜μ˜€μ§€λ§Œ μ΅œμ‹  λ²€μΉ˜λ§ˆν¬μ— λΉ„ν•˜μ—¬ 더 λ§Žμ€ μΆ”λ‘  μ‹œκ°„μ΄ ν•„μš”ν•˜μ—¬ μ‹€μ‹œκ°„ μž‘μ—…μ—λŠ” μ μš©ν•˜κΈ°κ°€ 쉽지 μ•Šλ‹€. μ‹€μ œ μ˜μƒμ—λŠ” λ‹¨μˆœν•œ μž‘μŒλ³΄λ‹€λŠ” μ˜μƒ νšλ“, μ €μž₯ λ“±κ³Ό 같은 ν”„λ‘œμ„ΈμŠ€λ₯Ό κ±°μΉ˜λ©΄μ„œ μ—¬λŸ¬ μš”μΈλ“€λ‘œ μΈν•œ λ‹€μ–‘ν•œ 작음, λΈ”λŸ¬μ™€ 같은 μ—΄ν™”κ°€ 혼재 λ˜μ–΄ μžˆλ‹€. μ‹€μ œ μž‘μŒμ— λŒ€ν•œ λ‹€μ–‘ν•œ κ°λ„μ—μ„œμ˜ 뢄석과 μ—¬λŸ¬ λͺ¨λΈλ§ μ‹€ν—˜, 그리고 μ˜μƒ 작음 및 λΈ”λŸ¬, μ••μΆ•κ³Ό 같은 볡합 λͺ¨λΈλ§μ΄ ν•„μš”ν•˜λ‹€. ν–₯ν›„μ—λŠ” μ΄λŸ¬ν•œ 점듀을 λ³΄μ™„ν•¨μœΌλ‘œμ¨ μ„±λŠ₯을 ν–₯μƒμ‹œν‚€κ³  λ„€νŠΈμ›Œν¬μ˜ 쑰정을 톡해 μ‹€μ‹œκ°„μœΌλ‘œ 적용될 수 μžˆμŒμ„ κΈ°λŒ€ν•œλ‹€.1 Introduction 1 2 Review on Image Denoising Methods 4 2.1 Image Noise Models 4 2.2 Traditional Denoising Methods 8 2.2.1 TV-based regularization 9 2.2.2 Non-local regularization 9 2.2.3 Sparse representation 10 2.2.4 Low-rank minimization 10 2.3 CNN-based Denoising Methods 11 2.3.1 DnCNN 11 2.3.2 FFDNet 12 2.3.3 WDnCNN 12 2.3.4 DHDN 13 3 Proposed models 15 3.1 Related Works 15 3.1.1 Residual learning 15 3.1.2 Dilated convolution 16 3.2 Proposed Network Architecture 17 4 Experiments 21 4.1 Training Details 21 4.2 Synthetic Noise Reduction 23 4.2.1 Set12 denoising 24 4.2.2 Kodak24 and BSD68 denoising 30 4.3 Real Noise Reduction 34 4.3.1 DnD test results 35 4.3.2 NTIRE 2020 real image denoising challenge 42 5 Conclusion and Future Works 46 Abstract (in Korean) 54Docto

    Progressive Joint Low-light Enhancement and Noise Removal for Raw Images

    Full text link
    Low-light imaging on mobile devices is typically challenging due to insufficient incident light coming through the relatively small aperture, resulting in a low signal-to-noise ratio. Most of the previous works on low-light image processing focus either only on a single task such as illumination adjustment, color enhancement, or noise removal; or on a joint illumination adjustment and denoising task that heavily relies on short-long exposure image pairs collected from specific camera models, and thus these approaches are less practical and generalizable in real-world settings where camera-specific joint enhancement and restoration is required. To tackle this problem, in this paper, we propose a low-light image processing framework that performs joint illumination adjustment, color enhancement, and denoising. Considering the difficulty in model-specific data collection and the ultra-high definition of the captured images, we design two branches: a coefficient estimation branch as well as a joint enhancement and denoising branch. The coefficient estimation branch works in a low-resolution space and predicts the coefficients for enhancement via bilateral learning, whereas the joint enhancement and denoising branch works in a full-resolution space and progressively performs joint enhancement and denoising. In contrast to existing methods, our framework does not need to recollect massive data when being adapted to another camera model, which significantly reduces the efforts required to fine-tune our approach for practical usage. Through extensive experiments, we demonstrate its great potential in real-world low-light imaging applications when compared with current state-of-the-art methods

    Wavelet subband-specific learning for low-dose computed tomography denoising

    Get PDF
    Deep neural networks have shown great improvements in low-dose computed tomography (CT) denoising. Early algorithms were primarily optimized to obtain an accurate image with low distortion between the denoised image and reference full-dose image at the cost of yielding an overly smoothed unrealistic CT image. Recent research has sought to preserve the fine details of denoised images with high perceptual quality, which has been accompanied by a decrease in objective quality due to a trade-off between perceptual quality and distortion. We pursue a network that can generate accurate and realistic CT images with high objective and perceptual quality within one network, achieving a better perception-distortion trade-off. To achieve this goal, we propose a stationary wavelet transform-assisted network employing the characteristics of high- and low-frequency domains of the wavelet transform and frequency subband-specific losses defined in the wavelet domain. We first introduce a stationary wavelet transform for the network training procedure. Then, we train the network using objective loss functions defined for high- and low-frequency domains to enhance the objective quality of the denoised CT image. With this network design, we train the network again after replacing the objective loss functions with perceptual loss functions in high- and low-frequency domains. As a result, we acquired denoised CT images with high perceptual quality using this strategy while minimizing the objective quality loss. We evaluated our algorithms on the phantom and clinical images, and the quantitative and qualitative results indicate that ours outperform the existing state-of-the-art algorithms in terms of objective and perceptual quality.ope
    corecore