8 research outputs found
Decay2Distill: Leveraging spatial perturbation and regularization for self-supervised image denoising
Unpaired image denoising has achieved promising development over the last few
years. Regardless of the performance, methods tend to heavily rely on
underlying noise properties or any assumption which is not always practical.
Alternatively, if we can ground the problem from a structural perspective
rather than noise statistics, we can achieve a more robust solution. with such
motivation, we propose a self-supervised denoising scheme that is unpaired and
relies on spatial degradation followed by a regularized refinement. Our method
shows considerable improvement over previous methods and exhibited consistent
performance over different data domains
Rethinking gradient weights' influence over saliency map estimation
Class activation map (CAM) helps to formulate saliency maps that aid in
interpreting the deep neural network's prediction. Gradient-based methods are
generally faster than other branches of vision interpretability and independent
of human guidance. The performance of CAM-like studies depends on the governing
model's layer response, and the influences of the gradients. Typical
gradient-oriented CAM studies rely on weighted aggregation for saliency map
estimation by projecting the gradient maps into single weight values, which may
lead to over generalized saliency map. To address this issue, we use a global
guidance map to rectify the weighted aggregation operation during saliency
estimation, where resultant interpretations are comparatively clean er and
instance-specific. We obtain the global guidance map by performing elementwise
multiplication between the feature maps and their corresponding gradient maps.
To validate our study, we compare the proposed study with eight different
saliency visualizers. In addition, we use seven commonly used evaluation
metrics for quantitative comparison. The proposed scheme achieves significant
improvement over the test images from the ImageNet, MS-COCO 14, and PASCAL VOC
2012 datasets
Denoising single images by feature ensemble revisited
Image denoising is still a challenging issue in many computer vision
sub-domains. Recent studies show that significant improvements are made
possible in a supervised setting. However, few challenges, such as spatial
fidelity and cartoon-like smoothing remain unresolved or decisively overlooked.
Our study proposes a simple yet efficient architecture for the denoising
problem that addresses the aforementioned issues. The proposed architecture
revisits the concept of modular concatenation instead of long and deeper
cascaded connections, to recover a cleaner approximation of the given image. We
find that different modules can capture versatile representations, and
concatenated representation creates a richer subspace for low-level image
restoration. The proposed architecture's number of parameters remains smaller
than the number for most of the previous networks and still achieves
significant improvements over the current state-of-the-art networks
SS-TTA : Test-Time Adaption for Self-Supervised Denoising Methods
Even though image denoising has already been studied for decades, recent progress in deep learning has provided novel and considerably better results for this classical signal reconstruction problem. One of the most significant advances in recent years has been relaxing the requirement of having noise-free (clean) images in the training dataset. By leveraging self-supervised learning, recent methods already reach the reconstruction quality of classical and some supervised schemes. In this paper, we propose SS-TTA, a generic test-time adaptation policy that can be applied on top of various self-supervised denoising methods. Taking a pre-trained self-supervised denoising model and a test image as input, our SS-TTA algorithm improves the denoising performance through a proposed ’inference-guided regularization’ process. Based on experiments with three synthetic and three real noise datasets, SS-TTA improves the denoising results of several state-of-the-art self-supervised methods, outperforms recent test-time adaptation approaches, and shows promising performance with supervised models. Finally, SS-TTA also generalizes to cases where the test-time noise distribution differs from the noise distribution of training images.©2023 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.fi=vertaisarvioitu|en=peerReviewed
Fast Single-Image HDR Tone-Mapping by Avoiding Base Layer Extraction
The tone-mapping algorithm compresses the high dynamic range (HDR) information into the standard dynamic range for regular devices. An ideal tone-mapping algorithm reproduces the HDR image without losing any vital information. The usual tone-mapping algorithms mostly deal with detail layer enhancement and gradient-domain manipulation with the help of a smoothing operator. However, these approaches often have to face challenges with over enhancement, halo effects, and over-saturation effects. To address these challenges, we propose a two-step solution to perform a tone-mapping operation using contrast enhancement. Our method improves the performance of the camera response model by utilizing the improved adaptive parameter selection and weight matrix extraction. Experiments show that our method performs reasonably well for overexposed and underexposed HDR images without producing any ringing or halo effects
Semi-supervised atmospheric component learning in low-light image problem.
Ambient lighting conditions play a crucial role in determining the perceptual quality of images from photographic devices. In general, inadequate transmission light and undesired atmospheric conditions jointly degrade the image quality. If we know the desired ambient factors associated with the given low-light image, we can recover the enhanced image easily. Typical deep networks perform enhancement mappings without investigating the light distribution and color formulation properties. This leads to a lack of image instance-adaptive performance in practice. On the other hand, physical model-driven schemes suffer from the need for inherent decompositions and multiple objective minimizations. Moreover, the above approaches are rarely data efficient or free of postprediction tuning. Influenced by the above issues, this study presents a semisupervised training method using no-reference image quality metrics for low-light image restoration. We incorporate the classical haze distribution model to explore the physical properties of the given image to learn the effect of atmospheric components and minimize a single objective for restoration. We validate the performance of our network for six widely used low-light datasets. Experimental studies show that our proposed study achieves a competitive performance for no-reference metrics compared to current state-of-the-art methods. We also show the improved generalization performance of our proposed method which is efficient in preserving face identities in extreme low-light scenarios
Semi-supervised atmospheric component learning in low-light image problem
Ambient lighting conditions play a crucial role in determining the perceptual quality of images from photographic devices. In general, inadequate transmission light and undesired atmospheric conditions jointly degrade the image quality. If we know the desired ambient factors associated with the given low-light image, we can recover the enhanced image easily. Typical deep networks perform enhancement mappings without investigating the light distribution and color formulation properties. This leads to a lack of image instance-adaptive performance in practice. On the other hand, physical model-driven schemes suffer from the need for inherent decompositions and multiple objective minimizations. Moreover, the above approaches are rarely data efficient or free of postprediction tuning. Influenced by the above issues, this study presents a semisupervised training method using no-reference image quality metrics for low-light image restoration. We incorporate the classical haze distribution model to explore the physical properties of the given image to learn the effect of atmospheric components and minimize a single objective for restoration. We validate the performance of our network for six widely used low-light datasets. Experimental studies show that our proposed study achieves a competitive performance for no-reference metrics compared to current state-of-the-art methods. We also show the improved generalization performance of our proposed method which is efficient in preserving face identities in extreme low-light scenarios
Rethinking Gradient Weight’s Influence over Saliency Map Estimation
Class activation map (CAM) helps to formulate saliency maps that aid in interpreting the deep neural network’s prediction. Gradient-based methods are generally faster than other branches of vision interpretability and independent of human guidance. The performance of CAM-like studies depends on the governing model’s layer response and the influences of the gradients. Typical gradient-oriented CAM studies rely on weighted aggregation for saliency map estimation by projecting the gradient maps into single-weight values, which may lead to an over-generalized saliency map. To address this issue, we use a global guidance map to rectify the weighted aggregation operation during saliency estimation, where resultant interpretations are comparatively cleaner and instance-specific. We obtain the global guidance map by performing elementwise multiplication between the feature maps and their corresponding gradient maps. To validate our study, we compare the proposed study with nine different saliency visualizers. In addition, we use seven commonly used evaluation metrics for quantitative comparison. The proposed scheme achieves significant improvement over the test images from the ImageNet, MS-COCO 14, and PASCAL VOC 2012 datasets