3,276 research outputs found
A Survey of Deep Face Restoration: Denoise, Super-Resolution, Deblur, Artifact Removal
Face Restoration (FR) aims to restore High-Quality (HQ) faces from
Low-Quality (LQ) input images, which is a domain-specific image restoration
problem in the low-level computer vision area. The early face restoration
methods mainly use statistic priors and degradation models, which are difficult
to meet the requirements of real-world applications in practice. In recent
years, face restoration has witnessed great progress after stepping into the
deep learning era. However, there are few works to study deep learning-based
face restoration methods systematically. Thus, this paper comprehensively
surveys recent advances in deep learning techniques for face restoration.
Specifically, we first summarize different problem formulations and analyze the
characteristic of the face image. Second, we discuss the challenges of face
restoration. Concerning these challenges, we present a comprehensive review of
existing FR methods, including prior based methods and deep learning-based
methods. Then, we explore developed techniques in the task of FR covering
network architectures, loss functions, and benchmark datasets. We also conduct
a systematic benchmark evaluation on representative methods. Finally, we
discuss future directions, including network designs, metrics, benchmark
datasets, applications,etc. We also provide an open-source repository for all
the discussed methods, which is available at
https://github.com/TaoWangzj/Awesome-Face-Restoration.Comment: 21 pages, 19 figure
Recommended from our members
Variational Multi-Task Models for Image Analysis: Applications to Magnetic Resonance Imaging
This thesis deals with the study and development of several variational multi-task models for solving inverse problems in imaging, with a particular focus on Magnetic Resonance Imaging (MRI). In most image processing problems, one usually deals with the reconstruction task, i.e., the task of reconstructing an image from indirect measurements, and then performs various operations, one after the other (i.e. sequentially), to improve the quality of the reconstruction and to extract useful information.
However, recent developments in a variational context, have shown that performing those tasks jointly (i.e. in a multi-task framework) offers great benefits, and this is the perspective that we follow in this thesis. We go beyond traditional sequential approaches and set a new basis for variational multi-task methods for MRI analysis. We demonstrate that by sharing representation between tasks and carefully interconnecting them, one can create synergies across challenging problems and reduce error propagation.
More precisely, firstly we propose a multi-task variational model to tackle the problems of image reconstruction and image segmentation using non-convex Bregman iteration. We describe theoretical and numerical details of the problem and its optimisation scheme. Moreover, we show that our multi-task model achieves better results in several examples and MRI applications than existing approaches in the same context.
Secondly, we show that our approach can be extended to a multi-task reconstruction and segmentation model for the nonlinear inverse problem of velocity-encoded MRI. In this context, the aim is to estimate not only the magnitude from MRI data, but also the phase and its flow information, whilst simultaneously identify regions of interest through the segmentation task.
Finally, we go beyond two-task frameworks and introduce for the first time a variational multi-task model to handle three imaging tasks. To this end, we design a variational multi-task framework addressing reconstruction, super-resolution and registration for improving the quality of MRI reconstruction. We demonstrate that our model is theoretically well-motivated and it outperforms sequential models whilst requiring less computational cost. Furthermore, we show through experimental results the potential of this approach for clinical applications
Tomographic neurofeedback : a new technique for the self-regulation of brain electrical activity
A major limitation of the current neurofeedback paradigm is the limited information provided by a single or a small number of electrodes placed on the scalp. A considerable improvement of the neurofeedback efficacy and specificity could be obtained feeding back brain activity of delimited structures. While traditional EEG information reflects the superposition of the electrical activity of a large number of neurons, by means of inverse solutions such as the Low-Resolution Electromagnetic Tomography (LORETA) spatially delimited brain activity can be evaluated in neocortical tissue. In this Dissertation we implement LORETA neurofeedback, we introduce a new feedback function ( 1 ) sensitive to dynamic change over time, and we clarify several issues related to the learning process observable with neurofeedback. The reported set of three experiments is the first attempt I am aware of to prove learning of brain current density activity. Three individuals were trained to improve brain activation (suppress low Alpha (8-10 Hz) and enhance low Beta (16 -20 Hz) current density) in the anterior cingulate gyros cognitive division (ACcd). Participants took part of six experimental sessions, each lasting approximately 30 minutes. Randomization-Permutation ANCOVA tests were conducted on recordings of the neurofeedback training. In addition a randomized trial was performed at the end of the treatment. During eight two-minutes periods (trials) participants were asked to try to obtain as many rewards as they could ( 4 1 trials) or as few rewards as they could ( 4 0 trials). The order of trials was decided at random. The hypothesis under testing was that participants acquired volitional control over their brain activity so to be able to obtain more rewards during the plus condition as compared to the minus condition. We found evidence of volitional control for two subjects (p=0.043 and p=O.l) and no evidence of volitional control for one of them (p=0.27 1). The combination of the three p-values provided an overall probability value for this experiment of 0.012 with the additive method and 0.035 with the multiplicative method. These results strongly support the hypothesis of volitional control across the experimental group. Trends of the Beta/ Alpha power ratio in the ACcd were in the expected direction for all the three subjects, however the combined p-values did not reach significance. With as few as six training sessions, typically insufficient to produce any form of learning with scalp neurofeedback, the experiment showed overall signs of volitional control of the electrical activity of the ACcd. Possible applications of the technique are important and include the treatment of epileptic foci, the treatment of specific brain regions damaged as a consequence of traumatic brain injury, and in general of any specific cortical electrical activity
Camera-independent learning and image quality assessment for super-resolution
An increasing number of applications require high-resolution images in situations where the access to the sensor and the knowledge of its specifications are limited. In this thesis, the problem of blind super-resolution is addressed, here defined as the estimation of a high-resolution image from one or more low-resolution inputs, under the condition that the degradation model parameters are unknown. The assessment of super-resolved results, using objective measures of image quality, is also addressed.Learning-based methods have been successfully applied to the single frame super-resolution problem in the past. However, sensor characteristics such as the Point Spread Function (PSF) must often be known. In this thesis, a learning-based approach is adapted to work without the knowledge of the PSF thus making the framework camera-independent. However, the goal is not only to super-resolve an image under this limitation, but also to provide an estimation of the best PSF, consisting of a theoretical model with one unknown parameter.In particular, two extensions of a method performing belief propagation on a Markov Random Field are presented. The first method finds the best PSF parameter by performing a search for the minimum mean distance between training examples and patches from the input image. In the second method, the best PSF parameter and the super-resolution result are found simultaneously by providing a range of possible PSF parameters from which the super-resolution algorithm will choose from. For both methods, a first estimate is obtained through blind deconvolution and an uncertainty is calculated in order to restrict the search.Both camera-independent adaptations are compared and analyzed in various experiments, and a set of key parameters are varied to determine their effect on both the super-resolution and the PSF parameter recovery results. The use of quality measures is thus essential to quantify the improvements obtained from the algorithms. A set of measures is chosen that represents different aspects of image quality: the signal fidelity, the perceptual quality and the localization and scale of the edges.Results indicate that both methods improve similarity to the ground truth and can in general refine the initial PSF parameter estimate towards the true value. Furthermore, the similarity measure results show that the chosen learning-based framework consistently improves a measure designed for perceptual quality
Coherent correlation imaging for resolving fluctuating states of matter
Fluctuations and stochastic transitions are ubiquitous in nanometre-scale systems, especially in the presence of disorder. However, their direct observation has so far been impeded by a seemingly fundamental, signal-limited compromise between spatial and temporal resolution. Here we develop coherent correlation imaging (CCI) to overcome this dilemma. Our method begins by classifying recorded camera frames in Fourier space. Contrast and spatial resolution emerge by averaging selectively over same-state frames. Temporal resolution down to the acquisition time of a single frame arises independently from an exceptionally low misclassification rate, which we achieve by combining a correlation-based similarity metric1,2 with a modified, iterative hierarchical clustering algorithm3,4. We apply CCI to study previously inaccessible magnetic fluctuations in a highly degenerate magnetic stripe domain state with nanometre-scale resolution. We uncover an intricate network of transitions between more than 30 discrete states. Our spatiotemporal data enable us to reconstruct the pinning energy landscape and to thereby explain the dynamics observed on a microscopic level. CCI massively expands the potential of emerging high-coherence X-ray sources and paves the way for addressing large fundamental questions such as the contribution of pinning5–8 and topology9–12 in phase transitions and the role of spin and charge order fluctuations in high-temperature superconductivity13,14
An attention residual u-net with differential preprocessing and geometric postprocessing: Learning how to segment vasculature including intracranial aneurysms
Objective Intracranial aneurysms (IA) are lethal, with high morbidity and mortality rates. Reliable, rapid, and accurate segmentation of IAs and their adjacent vasculature from medical imaging data is important to improve the clinical management of patients with IAs. However, due to the blurred boundaries and complex structure of IAs and overlapping with brain tissue or other cerebral arteries, image segmentation of IAs remains challenging. This study aimed to develop an attention residual U-Net (ARU-Net) architecture with differential preprocessing and geometric postprocessing for automatic segmentation of IAs and their adjacent arteries in conjunction with 3D rotational angiography (3DRA) images. Methods The proposed ARU-Net followed the classic U-Net framework with the following key enhancements. First, we preprocessed the 3DRA images based on boundary enhancement to capture more contour information and enhance the presence of small vessels. Second, we introduced the long skip connections of the attention gate at each layer of the fully convolutional decoder-encoder structure to emphasize the field of view (FOV) for IAs. Third, residual-based short skip connections were also embedded in each layer to implement in-depth supervision to help the network converge. Fourth, we devised a multiscale supervision strategy for independent prediction at different levels of the decoding path, integrating multiscale semantic information to facilitate the segmentation of small vessels. Fifth, the 3D conditional random field (3DCRF) and 3D connected component optimization (3DCCO) were exploited as postprocessing to optimize the segmentation results. Results Comprehensive experimental assessments validated the effectiveness of our ARU-Net. The proposed ARU-Net model achieved comparable or superior performance to the state-of-the-art methods through quantitative and qualitative evaluations. Notably, we found that ARU-Net improved the identification of arteries connecting to an IA, including small arteries that were hard to recognize by other methods. Consequently, IA geometries segmented by the proposed ARU-Net model yielded superior performance during subsequent computational hemodynamic studies (also known as patient-specific computational fluid dynamics [CFD] simulations). Furthermore, in an ablation study, the five key enhancements mentioned above were confirmed. Conclusions The proposed ARU-Net model can automatically segment the IAs in 3DRA images with relatively high accuracy and potentially has significant value for clinical computational hemodynamic analysis
- …