25,376 research outputs found
Domain Fingerprints for No-reference Image Quality Assessment
Human fingerprints are detailed and nearly unique markers of human identity.
Such a unique and stable fingerprint is also left on each acquired image. It
can reveal how an image was degraded during the image acquisition procedure and
thus is closely related to the quality of an image. In this work, we propose a
new no-reference image quality assessment (NR-IQA) approach called domain-aware
IQA (DA-IQA), which for the first time introduces the concept of domain
fingerprint to the NR-IQA field. The domain fingerprint of an image is learned
from image collections of different degradations and then used as the unique
characteristics to identify the degradation sources and assess the quality of
the image. To this end, we design a new domain-aware architecture, which
enables simultaneous determination of both the distortion sources and the
quality of an image. With the distortion in an image better characterized, the
image quality can be more accurately assessed, as verified by extensive
experiments, which show that the proposed DA-IQA performs better than almost
all the compared state-of-the-art NR-IQA methods.Comment: accepted by IEEE Transactions on Circuits and Systems for Video
Technology (TCSVT
FaceQnet: Quality Assessment for Face Recognition based on Deep Learning
In this paper we develop a Quality Assessment approach for face recognition
based on deep learning. The method consists of a Convolutional Neural Network,
FaceQnet, that is used to predict the suitability of a specific input image for
face recognition purposes. The training of FaceQnet is done using the VGGFace2
database. We employ the BioLab-ICAO framework for labeling the VGGFace2 images
with quality information related to their ICAO compliance level. The
groundtruth quality labels are obtained using FaceNet to generate comparison
scores. We employ the groundtruth data to fine-tune a ResNet-based CNN, making
it capable of returning a numerical quality measure for each input image.
Finally, we verify if the FaceQnet scores are suitable to predict the expected
performance when employing a specific image for face recognition with a COTS
face recognition system. Several conclusions can be drawn from this work, most
notably: 1) we managed to employ an existing ICAO compliance framework and a
pretrained CNN to automatically label data with quality information, 2) we
trained FaceQnet for quality estimation by fine-tuning a pre-trained face
recognition network (ResNet-50), and 3) we have shown that the predictions from
FaceQnet are highly correlated with the face recognition accuracy of a
state-of-the-art commercial system not used during development. FaceQnet is
publicly available in GitHub.Comment: Preprint version of a paper accepted at ICB 201
- …