2,948 research outputs found
Balancing Biases and Preserving Privacy on Balanced Faces in the Wild
Demographic biases exist in current models used for facial recognition (FR).
Our Balanced Faces in the Wild (BFW) dataset is a proxy to measure bias across
ethnicity and gender subgroups, allowing one to characterize FR performances
per subgroup. We show that results are non-optimal when a single score
threshold determines whether sample pairs are genuine or imposters.
Furthermore, within subgroups, performance often varies significantly from the
global average. Thus, specific error rates only hold for populations matching
the validation data. We mitigate the imbalanced performances using a novel
domain adaptation learning scheme on the facial features extracted from
state-of-the-art neural networks, boosting the average performance. The
proposed method also preserves identity information while removing demographic
knowledge. The removal of demographic knowledge prevents potential biases from
being injected into decision-making and protects privacy since demographic
information is no longer available. We explore the proposed method and show
that subgroup classifiers can no longer learn from the features projected using
our domain adaptation scheme. For source code and data, see
https://github.com/visionjo/facerec-bias-bfw.Comment: arXiv admin note: text overlap with arXiv:2102.0894
Beyond Accuracy: A Critical Review of Fairness in Machine Learning for Mobile and Wearable Computing
The field of mobile, wearable, and ubiquitous computing (UbiComp) is
undergoing a revolutionary integration of machine learning. Devices can now
diagnose diseases, predict heart irregularities, and unlock the full potential
of human cognition. However, the underlying algorithms are not immune to biases
with respect to sensitive attributes (e.g., gender, race), leading to
discriminatory outcomes. The research communities of HCI and AI-Ethics have
recently started to explore ways of reporting information about datasets to
surface and, eventually, counter those biases. The goal of this work is to
explore the extent to which the UbiComp community has adopted such ways of
reporting and highlight potential shortcomings. Through a systematic review of
papers published in the Proceedings of the ACM Interactive, Mobile, Wearable
and Ubiquitous Technologies (IMWUT) journal over the past 5 years (2018-2022),
we found that progress on algorithmic fairness within the UbiComp community
lags behind. Our findings show that only a small portion (5%) of published
papers adheres to modern fairness reporting, while the overwhelming majority
thereof focuses on accuracy or error metrics. In light of these findings, our
work provides practical guidelines for the design and development of ubiquitous
technologies that not only strive for accuracy but also for fairness
The State of AI Ethics Report (June 2020)
These past few months have been especially challenging, and the deployment of technology in ways hitherto untested at an unrivalled pace has left the internet and technology watchers aghast. Artificial intelligence has become the byword for technological progress and is being used in everything from helping us combat the COVID-19 pandemic to nudging our attention in different directions as we all spend increasingly larger amounts of time online. It has never been more important that we keep a sharp eye out on the development of this field and how it is shaping our society and interactions with each other. With this inaugural edition of the State of AI Ethics we hope to bring forward the most important developments that caught our attention at the Montreal AI Ethics Institute this past quarter. Our goal is to help you navigate this ever-evolving field swiftly and allow you and your organization to make informed decisions. This pulse-check for the state of discourse, research, and development is geared towards researchers and practitioners alike who are making decisions on behalf of their organizations in considering the societal impacts of AI-enabled solutions. We cover a wide set of areas in this report spanning Agency and Responsibility, Security and Risk, Disinformation, Jobs and Labor, the Future of AI Ethics, and more. Our staff has worked tirelessly over the past quarter surfacing signal from the noise so that you are equipped with the right tools and knowledge to confidently tread this complex yet consequential domain
A Survey of Face Recognition
Recent years witnessed the breakthrough of face recognition with deep
convolutional neural networks. Dozens of papers in the field of FR are
published every year. Some of them were applied in the industrial community and
played an important role in human life such as device unlock, mobile payment,
and so on. This paper provides an introduction to face recognition, including
its history, pipeline, algorithms based on conventional manually designed
features or deep learning, mainstream training, evaluation datasets, and
related applications. We have analyzed and compared state-of-the-art works as
many as possible, and also carefully designed a set of experiments to find the
effect of backbone size and data distribution. This survey is a material of the
tutorial named The Practical Face Recognition Technology in the Industrial
World in the FG2023
Survey of Social Bias in Vision-Language Models
In recent years, the rapid advancement of machine learning (ML) models,
particularly transformer-based pre-trained models, has revolutionized Natural
Language Processing (NLP) and Computer Vision (CV) fields. However, researchers
have discovered that these models can inadvertently capture and reinforce
social biases present in their training datasets, leading to potential social
harms, such as uneven resource allocation and unfair representation of specific
social groups. Addressing these biases and ensuring fairness in artificial
intelligence (AI) systems has become a critical concern in the ML community.
The recent introduction of pre-trained vision-and-language (VL) models in the
emerging multimodal field demands attention to the potential social biases
present in these models as well. Although VL models are susceptible to social
bias, there is a limited understanding compared to the extensive discussions on
bias in NLP and CV. This survey aims to provide researchers with a high-level
insight into the similarities and differences of social bias studies in
pre-trained models across NLP, CV, and VL. By examining these perspectives, the
survey aims to offer valuable guidelines on how to approach and mitigate social
bias in both unimodal and multimodal settings. The findings and recommendations
presented here can benefit the ML community, fostering the development of
fairer and non-biased AI models in various applications and research endeavors
Responsible and Regulatory Conform Machine Learning for Medicine: A Survey of Challenges and Solutions
Machine learning is expected to fuel significant improvements in medical
care. To ensure that fundamental principles such as beneficence, respect for
human autonomy, prevention of harm, justice, privacy, and transparency are
respected, medical machine learning systems must be developed responsibly. Many
high-level declarations of ethical principles have been put forth for this
purpose, but there is a severe lack of technical guidelines explicating the
practical consequences for medical machine learning. Similarly, there is
currently considerable uncertainty regarding the exact regulatory requirements
placed upon medical machine learning systems. This survey provides an overview
of the technical and procedural challenges involved in creating medical machine
learning systems responsibly and in conformity with existing regulations, as
well as possible solutions to address these challenges. First, a brief review
of existing regulations affecting medical machine learning is provided, showing
that properties such as safety, robustness, reliability, privacy, security,
transparency, explainability, and nondiscrimination are all demanded already by
existing law and regulations - albeit, in many cases, to an uncertain degree.
Next, the key technical obstacles to achieving these desirable properties are
discussed, as well as important techniques to overcome these obstacles in the
medical context. We notice that distribution shift, spurious correlations,
model underspecification, uncertainty quantification, and data scarcity
represent severe challenges in the medical context. Promising solution
approaches include the use of large and representative datasets and federated
learning as a means to that end, the careful exploitation of domain knowledge,
the use of inherently transparent models, comprehensive out-of-distribution
model testing and verification, as well as algorithmic impact assessments
- …