211,348 research outputs found
Conditional Image Synthesis by Generative Adversarial Modeling
Recent years, image synthesis has attracted more interests. This work explores the recovery of details (low-level information) from high-level features. The generative adversarial nets (GAN) has led to the explosion of image synthesis. Moving away from those application-oriented alternatives, this work investigates its intrinsic drawbacks and derives corresponding improvements in a theoretical manner.Based on GAN, this work further investigates the conditional image synthesis by incorporating an autoencoder (AE) to GAN. The GAN+AE structure has been demonstrated to be an effective framework for image manipulation. This work emphasizes the effectiveness of GAN+AE structure by proposing the conditional adversarial autoencoder (CAAE) for human facial age progression and regression. Instead of editing on the image level, i.e., explicitly changing the shape of face, adding wrinkle, etc., this work edits the high-level features which implicitly guide the recovery of images towards expected appearance.While GAN+AE being prevalent in image manipulation, its drawbacks lack exploration. For example, GAN+AE requires a weight to balance the effects of GAN and AE. An inappropriate weight would generate unstable results. This work provides an insight to such instability, which is due to the interaction between GAN and AE. Therefore, this work proposes the decoupled learning (GAN//AE) to avoid the interaction between them and achieve a robust and effective framework for image synthesis. Most existing works used GAN+AE structure could be easily adapted to the proposed GAN//AE structure to boost their robustness. Experimental results demonstrate the correctness and effectiveness of the provided derivation and proposed methods, respectively.In addition, this work extends the conditional image synthesis to the traditional area of image super-resolution, which recovers the high-resolution image according the low-resolution counterpart. Diverting from such traditional routine, this work explores a new research direction | reference-conditioned super-resolution, in which a reference image containing desired high-resolution texture details is used besides the low-resolution image. We focus on transferring the high-resolution texture from reference images to the super-resolution process without the constraint of content similarity between reference and target images, which is a key difference from previous example-based methods
Real-world super-resolution of face-images from surveillance cameras
Most existing face image Super-Resolution (SR) methods assume that the
Low-Resolution (LR) images were artificially downsampled from High-Resolution
(HR) images with bicubic interpolation. This operation changes the natural
image characteristics and reduces noise. Hence, SR methods trained on such data
most often fail to produce good results when applied to real LR images. To
solve this problem, we propose a novel framework for generation of realistic
LR/HR training pairs. Our framework estimates realistic blur kernels, noise
distributions, and JPEG compression artifacts to generate LR images with
similar image characteristics as the ones in the source domain. This allows us
to train a SR model using high quality face images as Ground-Truth (GT). For
better perceptual quality we use a Generative Adversarial Network (GAN) based
SR model where we have exchanged the commonly used VGG-loss [24] with
LPIPS-loss [52]. Experimental results on both real and artificially corrupted
face images show that our method results in more detailed reconstructions with
less noise compared to existing State-of-the-Art (SoTA) methods. In addition,
we show that the traditional non-reference Image Quality Assessment (IQA)
methods fail to capture this improvement and demonstrate that the more recent
NIMA metric [16] correlates better with human perception via Mean Opinion Rank
(MOR)
Recommended from our members
Face image super-resolution using 2D CCA
In this paper a face super-resolution method using two-dimensional canonical correlation analysis (2D CCA) is presented. A detail compensation step is followed to add high-frequency components to the reconstructed high-resolution face. Unlike most of the previous researches on face super-resolution algorithms that first transform the images into vectors, in our approach the relationship between the high-resolution and the low-resolution face image are maintained in their original 2D representation. In addition, rather than approximating the entire face, different parts of a face image are super-resolved separately to better preserve the local structure. The proposed method is compared with various state-of-the-art super-resolution algorithms using multiple evaluation criteria including face recognition performance. Results on publicly available datasets show that the proposed method super-resolves high quality face images which are very close to the ground-truth and performance gain is not dataset dependent. The method is very efficient in both the training and testing phases compared to the other approaches. © 2013 Elsevier B.V
- …