7,844 research outputs found
How Does the Low-Rank Matrix Decomposition Help Internal and External Learnings for Super-Resolution
Wisely utilizing the internal and external learning methods is a new
challenge in super-resolution problem. To address this issue, we analyze the
attributes of two methodologies and find two observations of their recovered
details: 1) they are complementary in both feature space and image plane, 2)
they distribute sparsely in the spatial space. These inspire us to propose a
low-rank solution which effectively integrates two learning methods and then
achieves a superior result. To fit this solution, the internal learning method
and the external learning method are tailored to produce multiple preliminary
results. Our theoretical analysis and experiment prove that the proposed
low-rank solution does not require massive inputs to guarantee the performance,
and thereby simplifying the design of two learning methods for the solution.
Intensive experiments show the proposed solution improves the single learning
method in both qualitative and quantitative assessments. Surprisingly, it shows
more superior capability on noisy images and outperforms state-of-the-art
methods
FSRNet: End-to-End Learning Face Super-Resolution with Facial Priors
Face Super-Resolution (SR) is a domain-specific super-resolution problem. The
specific facial prior knowledge could be leveraged for better super-resolving
face images. We present a novel deep end-to-end trainable Face Super-Resolution
Network (FSRNet), which makes full use of the geometry prior, i.e., facial
landmark heatmaps and parsing maps, to super-resolve very low-resolution (LR)
face images without well-aligned requirement. Specifically, we first construct
a coarse SR network to recover a coarse high-resolution (HR) image. Then, the
coarse HR image is sent to two branches: a fine SR encoder and a prior
information estimation network, which extracts the image features, and
estimates landmark heatmaps/parsing maps respectively. Both image features and
prior information are sent to a fine SR decoder to recover the HR image. To
further generate realistic faces, we propose the Face Super-Resolution
Generative Adversarial Network (FSRGAN) to incorporate the adversarial loss
into FSRNet. Moreover, we introduce two related tasks, face alignment and
parsing, as the new evaluation metrics for face SR, which address the
inconsistency of classic metrics w.r.t. visual perception. Extensive benchmark
experiments show that FSRNet and FSRGAN significantly outperforms state of the
arts for very LR face SR, both quantitatively and qualitatively. Code will be
made available upon publication.Comment: Chen and Tai contributed equally to this pape
Multiple Exemplars-based Hallucinationfor Face Super-resolution and Editing
Given a really low-resolution input image of a face (say 16x16 or 8x8
pixels), the goal of this paper is to reconstruct a high-resolution version
thereof. This, by itself, is an ill-posed problem, as the high-frequency
information is missing in the low-resolution input and needs to be
hallucinated, based on prior knowledge about the image content. Rather than
relying on a generic face prior, in this paper, we explore the use of a set of
exemplars, i.e. other high-resolution images of the same person. These guide
the neural network as we condition the output on them. Multiple exemplars work
better than a single one. To combine the information from multiple exemplars
effectively, we introduce a pixel-wise weight generation module. Besides
standard face super-resolution, our method allows to perform subtle face
editing simply by replacing the exemplars with another set with different
facial features. A user study is conducted and shows the super-resolved images
can hardly be distinguished from real images on the CelebA dataset. A
qualitative comparison indicates our model outperforms methods proposed in the
literature on the CelebA and WebFace dataset.Comment: accepted in ACCV 202
Face Hallucination via Deep Neural Networks.
We firstly address aligned low-resolution (LR) face images (i.e. 16X16 pixels) by designing a discriminative generative network, named URDGN. URDGN is composed of two networks: a generative model and a discriminative model.
We introduce a pixel-wise L2 regularization term to the generative model and exploit the feedback of the discriminative network to make the upsampled face images more similar to real ones.
We present an end-to-end transformative discriminative neural network (TDN) devised for super-resolving unaligned tiny face images. TDN embeds spatial transformation layers to enforce local receptive fields to line-up with similar spatial supports. To upsample noisy unaligned LR face images, we propose decoder-encoder-decoder networks. A transformative discriminative decoder network is employed to upsample and denoise LR inputs simultaneously. Then we project the intermediate HR faces to aligned and noise-free LR faces by a transformative encoder network. Finally, high-quality hallucinated HR images are generated by our second decoder. Furthermore, we present an end-to-end multiscale transformative discriminative neural network (MTDN) to super-resolve unaligned LR face images of different resolutions in a unified framework.
We propose a method that explicitly incorporates structural information of faces into the face super-resolution process by using a multi-task convolutional neural network (CNN). Our method not only uses low-level information (i.e. intensity similarity), but also middle-level information (i.e. face structure) to further explore spatial constraints of facial components from LR inputs images.
We demonstrate that supplementing residual images or feature maps with additional facial attribute information can significantly reduce the ambiguity in face super-resolution. To explore this idea, we develop an attribute-embedded upsampling network. In this manner, our method is able to super-resolve LR faces by a large upscaling factor while reducing the uncertainty of one-to-many mappings remarkably.
We further push the boundaries of hallucinating a tiny, non-frontal face image to understand how much of this is possible by leveraging the availability of large datasets and deep networks. To this end, we introduce a novel Transformative Adversarial Neural Network (TANN) to jointly frontalize very LR out-of-plane rotated face images (including profile views) and aggressively super-resolve them by 8X, regardless of their original poses and without using any 3D information. Besides recovering an HR face images from an LR version, this thesis also addresses the task of restoring realistic faces from stylized portrait images, which can also be regarded as face hallucination
- …