11,832 research outputs found
A Deep Face Identification Network Enhanced by Facial Attributes Prediction
In this paper, we propose a new deep framework which predicts facial
attributes and leverage it as a soft modality to improve face identification
performance. Our model is an end to end framework which consists of a
convolutional neural network (CNN) whose output is fanned out into two separate
branches; the first branch predicts facial attributes while the second branch
identifies face images. Contrary to the existing multi-task methods which only
use a shared CNN feature space to train these two tasks jointly, we fuse the
predicted attributes with the features from the face modality in order to
improve the face identification performance. Experimental results show that our
model brings benefits to both face identification as well as facial attribute
prediction performance, especially in the case of identity facial attributes
such as gender prediction. We tested our model on two standard datasets
annotated by identities and face attributes. Experimental results indicate that
the proposed model outperforms most of the current existing face identification
and attribute prediction methods
AnonymousNet: Natural Face De-Identification with Measurable Privacy
With billions of personal images being generated from social media and
cameras of all sorts on a daily basis, security and privacy are unprecedentedly
challenged. Although extensive attempts have been made, existing face image
de-identification techniques are either insufficient in photo-reality or
incapable of balancing privacy and usability qualitatively and quantitatively,
i.e., they fail to answer counterfactual questions such as "is it private
now?", "how private is it?", and "can it be more private?" In this paper, we
propose a novel framework called AnonymousNet, with an effort to address these
issues systematically, balance usability, and enhance privacy in a natural and
measurable manner. The framework encompasses four stages: facial attribute
estimation, privacy-metric-oriented face obfuscation, directed natural image
synthesis, and adversarial perturbation. Not only do we achieve the
state-of-the-arts in terms of image quality and attribute prediction accuracy,
we are also the first to show that facial privacy is measurable, can be
factorized, and accordingly be manipulated in a photo-realistic fashion to
fulfill different requirements and application scenarios. Experiments further
demonstrate the effectiveness of the proposed framework.Comment: CVPR-19 Workshop on Computer Vision: Challenges and Opportunities for
Privacy and Security (CV-COPS 2019
A Survey of Deep Facial Attribute Analysis
Facial attribute analysis has received considerable attention when deep
learning techniques made remarkable breakthroughs in this field over the past
few years. Deep learning based facial attribute analysis consists of two basic
sub-issues: facial attribute estimation (FAE), which recognizes whether facial
attributes are present in given images, and facial attribute manipulation
(FAM), which synthesizes or removes desired facial attributes. In this paper,
we provide a comprehensive survey of deep facial attribute analysis from the
perspectives of both estimation and manipulation. First, we summarize a general
pipeline that deep facial attribute analysis follows, which comprises two
stages: data preprocessing and model construction. Additionally, we introduce
the underlying theories of this two-stage pipeline for both FAE and FAM.
Second, the datasets and performance metrics commonly used in facial attribute
analysis are presented. Third, we create a taxonomy of state-of-the-art methods
and review deep FAE and FAM algorithms in detail. Furthermore, several
additional facial attribute related issues are introduced, as well as relevant
real-world applications. Finally, we discuss possible challenges and promising
future research directions.Comment: submitted to International Journal of Computer Vision (IJCV
Gender Privacy: An Ensemble of Semi Adversarial Networks for Confounding Arbitrary Gender Classifiers
Recent research has proposed the use of Semi Adversarial Networks (SAN) for
imparting privacy to face images. SANs are convolutional autoencoders that
perturb face images such that the perturbed images cannot be reliably used by
an attribute classifier (e.g., a gender classifier) but can still be used by a
face matcher for matching purposes. However, the generalizability of SANs
across multiple arbitrary gender classifiers has not been demonstrated in the
literature. In this work, we tackle the generalization issue by designing an
ensemble SAN model that generates a diverse set of perturbed outputs for a
given input face image. This is accomplished by enforcing diversity among the
individual models in the ensemble through the use of different data
augmentation techniques. The goal is to ensure that at least one of the
perturbed output faces will confound an arbitrary, previously unseen gender
classifier. Extensive experiments using different unseen gender classifiers and
face matchers are performed to demonstrate the efficacy of the proposed
paradigm in imparting gender privacy to face images.Comment: Published in Proc. of IEEE 9th International Conference on
Biometrics: Theory, Applications and Systems (BTAS), (Los Angeles, CA),
October 201
Using Deep Cross Modal Hashing and Error Correcting Codes for Improving the Efficiency of Attribute Guided Facial Image Retrieval
With benefits of fast query speed and low storage cost, hashing-based image
retrieval approaches have garnered considerable attention from the research
community. In this paper, we propose a novel Error-Corrected Deep Cross Modal
Hashing (CMH-ECC) method which uses a bitmap specifying the presence of certain
facial attributes as an input query to retrieve relevant face images from the
database. In this architecture, we generate compact hash codes using an
end-to-end deep learning module, which effectively captures the inherent
relationships between the face and attribute modality. We also integrate our
deep learning module with forward error correction codes to further reduce the
distance between different modalities of the same subject. Specifically, the
properties of deep hashing and forward error correction codes are exploited to
design a cross modal hashing framework with high retrieval performance.
Experimental results using two standard datasets with facial attributes-image
modalities indicate that our CMH-ECC face image retrieval model outperforms
most of the current attribute-based face image retrieval approaches.Comment: To be published in Proc. IEEE Global SIP 201
Semi-Latent GAN: Learning to generate and modify facial images from attributes
Generating and manipulating human facial images using high-level attributal
controls are important and interesting problems. The models proposed in
previous work can solve one of these two problems (generation or manipulation),
but not both coherently. This paper proposes a novel model that learns how to
both generate and modify the facial image from high-level semantic attributes.
Our key idea is to formulate a Semi-Latent Facial Attribute Space (SL-FAS) to
systematically learn relationship between user-defined and latent attributes,
as well as between those attributes and RGB imagery. As part of this newly
formulated space, we propose a new model --- SL-GAN which is a specific form of
Generative Adversarial Network. Finally, we present an iterative training
algorithm for SL-GAN. The experiments on recent CelebA and CASIA-WebFace
datasets validate the effectiveness of our proposed framework. We will also
make data, pre-trained models and code available.Comment: 10 pages, submitted to ICCV 201
Curriculum Learning of Visual Attribute Clusters for Multi-Task Classification
Visual attributes, from simple objects (e.g., backpacks, hats) to
soft-biometrics (e.g., gender, height, clothing) have proven to be a powerful
representational approach for many applications such as image description and
human identification. In this paper, we introduce a novel method to combine the
advantages of both multi-task and curriculum learning in a visual attribute
classification framework. Individual tasks are grouped after performing
hierarchical clustering based on their correlation. The clusters of tasks are
learned in a curriculum learning setup by transferring knowledge between
clusters. The learning process within each cluster is performed in a multi-task
classification setup. By leveraging the acquired knowledge, we speed-up the
process and improve performance. We demonstrate the effectiveness of our method
via ablation studies and a detailed analysis of the covariates, on a variety of
publicly available datasets of humans standing with their full-body visible.
Extensive experimentation has proven that the proposed approach boosts the
performance by 4% to 10%.Comment: Published in Pattern Recognitio
Learning from Longitudinal Face Demonstration - Where Tractable Deep Modeling Meets Inverse Reinforcement Learning
This paper presents a novel Subject-dependent Deep Aging Path (SDAP), which
inherits the merits of both Generative Probabilistic Modeling and Inverse
Reinforcement Learning to model the facial structures and the longitudinal face
aging process of a given subject. The proposed SDAP is optimized using
tractable log-likelihood objective functions with Convolutional Neural Networks
(CNNs) based deep feature extraction. Instead of applying a fixed aging
development path for all input faces and subjects, SDAP is able to provide the
most appropriate aging development path for individual subject that optimizes
the reward aging formulation. Unlike previous methods that can take only one
image as the input, SDAP further allows multiple images as inputs, i.e. all
information of a subject at either the same or different ages, to produce the
optimal aging path for the given subject. Finally, SDAP allows efficiently
synthesizing in-the-wild aging faces. The proposed model is experimented in
both tasks of face aging synthesis and cross-age face verification. The
experimental results consistently show SDAP achieves the state-of-the-art
performance on numerous face aging databases, i.e. FG-NET, MORPH, AginG Faces
in the Wild (AGFW), and Cross-Age Celebrity Dataset (CACD). Furthermore, we
also evaluate the performance of SDAP on large-scale Megaface challenge to
demonstrate the advantages of the proposed solution
cvpaper.challenge in 2016: Futuristic Computer Vision through 1,600 Papers Survey
The paper gives futuristic challenges disscussed in the cvpaper.challenge. In
2015 and 2016, we thoroughly study 1,600+ papers in several
conferences/journals such as CVPR/ICCV/ECCV/NIPS/PAMI/IJCV
cvpaper.challenge in 2015 - A review of CVPR2015 and DeepSurvey
The "cvpaper.challenge" is a group composed of members from AIST, Tokyo Denki
Univ. (TDU), and Univ. of Tsukuba that aims to systematically summarize papers
on computer vision, pattern recognition, and related fields. For this
particular review, we focused on reading the ALL 602 conference papers
presented at the CVPR2015, the premier annual computer vision event held in
June 2015, in order to grasp the trends in the field. Further, we are proposing
"DeepSurvey" as a mechanism embodying the entire process from the reading
through all the papers, the generation of ideas, and to the writing of paper.Comment: Survey Pape
- …