8,845 research outputs found
Analyzing and Reducing the Damage of Dataset Bias to Face Recognition With Synthetic Data
It is well known that deep learning approaches to facerecognition suffer from various biases in the available train-ing data. In this work, we demonstrate the large potentialof synthetic data for analyzing and reducing the negativeeffects of dataset bias on deep face recognition systems. Inparticular we explore two complementary application areasfor synthetic face images: 1) Using fully annotated syntheticface images we can study the face recognition rate as afunction of interpretable parameters such as face pose. Thisenables us to systematically analyze the effect of differenttypes of dataset biases on the generalization ability of neu-ral network architectures. Our analysis reveals that deeperneural network architectures can generalize better to un-seen face poses. Furthermore, our study shows that currentneural network architectures cannot disentangle face poseand facial identity, which limits their generalization ability.2) We pre-train neural networks with large-scale syntheticdata that is highly variable in face pose and the number offacial identities. After a subsequent fine-tuning with real-world data, we observe that the damage of dataset bias inthe real-world data is largely reduced. Furthermore, wedemonstrate that the size of real-world datasets can be re-duced by 75% while maintaining competitive face recogni-tion performance. The data and software used in this workare publicly available
High-Accuracy Facial Depth Models derived from 3D Synthetic Data
In this paper, we explore how synthetically generated 3D face models can be
used to construct a high accuracy ground truth for depth. This allows us to
train the Convolutional Neural Networks (CNN) to solve facial depth estimation
problems. These models provide sophisticated controls over image variations
including pose, illumination, facial expressions and camera position. 2D
training samples can be rendered from these models, typically in RGB format,
together with depth information. Using synthetic facial animations, a dynamic
facial expression or facial action data can be rendered for a sequence of image
frames together with ground truth depth and additional metadata such as head
pose, light direction, etc. The synthetic data is used to train a CNN based
facial depth estimation system which is validated on both synthetic and real
images. Potential fields of application include 3D reconstruction, driver
monitoring systems, robotic vision systems, and advanced scene understanding
Towards causal benchmarking of bias in face analysis algorithms
Measuring algorithmic bias is crucial both to assess algorithmic fairness, and to guide the improvement of algorithms. Current methods to measure algorithmic bias in computer vision, which are based on observational datasets, are inadequate for this task because they conflate algorithmic bias with dataset bias.
To address this problem we develop an experimental method for measuring algorithmic bias of face analysis algorithms, which manipulates directly the attributes of interest, e.g., gender and skin tone, in order to reveal causal links between attribute variation and performance change. Our proposed method is based on generating synthetic ``transects'' of matched sample images that are designed to differ along specific attributes while leaving other attributes constant. A crucial aspect of our approach is relying on the perception of human observers, both to guide manipulations, and to measure algorithmic bias.
Besides allowing the measurement of algorithmic bias, synthetic transects have other advantages with respect to observational datasets: they sample attributes more evenly allowing for more straightforward bias analysis on minority and intersectional groups, they enable prediction of bias in new scenarios, they greatly reduce ethical and legal challenges, and they are economical and fast to obtain, helping make bias testing affordable and widely available.
We validate our method by comparing it to a study that employs the traditional observational method for analyzing bias in gender classification algorithms. The two methods reach different conclusions. While the observational method reports gender and skin color biases, the experimental method reveals biases due to gender, hair length, age, and facial hair
Bias and unfairness in machine learning models: a systematic literature review
One of the difficulties of artificial intelligence is to ensure that model
decisions are fair and free of bias. In research, datasets, metrics,
techniques, and tools are applied to detect and mitigate algorithmic unfairness
and bias. This study aims to examine existing knowledge on bias and unfairness
in Machine Learning models, identifying mitigation methods, fairness metrics,
and supporting tools. A Systematic Literature Review found 40 eligible articles
published between 2017 and 2022 in the Scopus, IEEE Xplore, Web of Science, and
Google Scholar knowledge bases. The results show numerous bias and unfairness
detection and mitigation approaches for ML technologies, with clearly defined
metrics in the literature, and varied metrics can be highlighted. We recommend
further research to define the techniques and metrics that should be employed
in each case to standardize and ensure the impartiality of the machine learning
model, thus, allowing the most appropriate metric to detect bias and unfairness
in a given context
Simulated Adversarial Testing of Face Recognition Models
Most machine learning models are validated and tested on fixed datasets. This
can give an incomplete picture of the capabilities and weaknesses of the model.
Such weaknesses can be revealed at test time in the real world. The risks
involved in such failures can be loss of profits, loss of time or even loss of
life in certain critical applications. In order to alleviate this issue,
simulators can be controlled in a fine-grained manner using interpretable
parameters to explore the semantic image manifold. In this work, we propose a
framework for learning how to test machine learning algorithms using simulators
in an adversarial manner in order to find weaknesses in the model before
deploying it in critical scenarios. We apply this model in a face recognition
scenario. We are the first to show that weaknesses of models trained on real
data can be discovered using simulated samples. Using our proposed method, we
can find adversarial synthetic faces that fool contemporary face recognition
models. This demonstrates the fact that these models have weaknesses that are
not measured by commonly used validation datasets. We hypothesize that this
type of adversarial examples are not isolated, but usually lie in connected
components in the latent space of the simulator. We present a method to find
these adversarial regions as opposed to the typical adversarial points found in
the adversarial example literature
- …