2 research outputs found
Segmentation-Aware Image Denoising without Knowing True Segmentation
Several recent works discussed application-driven image restoration neural
networks, which are capable of not only removing noise in images but also
preserving their semantic-aware details, making them suitable for various
high-level computer vision tasks as the pre-processing step. However, such
approaches require extra annotations for their high-level vision tasks, in
order to train the joint pipeline using hybrid losses. The availability of
those annotations is yet often limited to a few image sets, potentially
restricting the general applicability of these methods to denoising more unseen
and unannotated images. Motivated by that, we propose a segmentation-aware
image denoising model dubbed U-SAID, based on a novel unsupervised approach
with a pixel-wise uncertainty loss. U-SAID does not need any ground-truth
segmentation map, and thus can be applied to any image dataset. It generates
denoised images with comparable or even better quality, and the denoised
results show stronger robustness for subsequent semantic segmentation tasks,
when compared to either its supervised counterpart or classical
"application-agnostic" denoisers. Moreover, we demonstrate the superior
generalizability of U-SAID in three-folds, by plugging its "universal" denoiser
without fine-tuning: (1) denoising unseen types of images; (2) denoising as
pre-processing for segmenting unseen noisy images; and (3) denoising for unseen
high-level tasks. Extensive experiments demonstrate the effectiveness,
robustness and generalizability of the proposed U-SAID over various popular
image sets
Recurrent Exposure Generation for Low-Light Face Detection
Face detection from low-light images is challenging due to limited photos and
inevitable noise, which, to make the task even harder, are often spatially
unevenly distributed. A natural solution is to borrow the idea from
multi-exposure, which captures multiple shots to obtain well-exposed images
under challenging conditions. High-quality implementation/approximation of
multi-exposure from a single image is however nontrivial. Fortunately, as shown
in this paper, neither is such high-quality necessary since our task is face
detection rather than image enhancement. Specifically, we propose a novel
Recurrent Exposure Generation (REG) module and couple it seamlessly with a
Multi-Exposure Detection (MED) module, and thus significantly improve face
detection performance by effectively inhibiting non-uniform illumination and
noise issues. REG produces progressively and efficiently intermediate images
corresponding to various exposure settings, and such pseudo-exposures are then
fused by MED to detect faces across different lighting conditions. The proposed
method, named REGDet, is the first `detection-with-enhancement' framework for
low-light face detection. It not only encourages rich interaction and feature
fusion across different illumination levels, but also enables effective
end-to-end learning of the REG component to be better tailored for face
detection. Moreover, as clearly shown in our experiments, REG can be flexibly
coupled with different face detectors without extra low/normal-light image
pairs for training. We tested REGDet on the DARK FACE low-light face benchmark
with thorough ablation study, where REGDet outperforms previous
state-of-the-arts by a significant margin, with only negligible extra
parameters.Comment: 11 page