1,562 research outputs found
Fairness in Recommendation: Foundations, Methods and Applications
As one of the most pervasive applications of machine learning, recommender
systems are playing an important role on assisting human decision making. The
satisfaction of users and the interests of platforms are closely related to the
quality of the generated recommendation results. However, as a highly
data-driven system, recommender system could be affected by data or algorithmic
bias and thus generate unfair results, which could weaken the reliance of the
systems. As a result, it is crucial to address the potential unfairness
problems in recommendation settings. Recently, there has been growing attention
on fairness considerations in recommender systems with more and more literature
on approaches to promote fairness in recommendation. However, the studies are
rather fragmented and lack a systematic organization, thus making it difficult
to penetrate for new researchers to the domain. This motivates us to provide a
systematic survey of existing works on fairness in recommendation. This survey
focuses on the foundations for fairness in recommendation literature. It first
presents a brief introduction about fairness in basic machine learning tasks
such as classification and ranking in order to provide a general overview of
fairness research, as well as introduce the more complex situations and
challenges that need to be considered when studying fairness in recommender
systems. After that, the survey will introduce fairness in recommendation with
a focus on the taxonomies of current fairness definitions, the typical
techniques for improving fairness, as well as the datasets for fairness studies
in recommendation. The survey also talks about the challenges and opportunities
in fairness research with the hope of promoting the fair recommendation
research area and beyond.Comment: Accepted by ACM Transactions on Intelligent Systems and Technology
(TIST
A Survey on Fairness-aware Recommender Systems
As information filtering services, recommender systems have extremely
enriched our daily life by providing personalized suggestions and facilitating
people in decision-making, which makes them vital and indispensable to human
society in the information era. However, as people become more dependent on
them, recent studies show that recommender systems potentially own
unintentional impacts on society and individuals because of their unfairness
(e.g., gender discrimination in job recommendations). To develop trustworthy
services, it is crucial to devise fairness-aware recommender systems that can
mitigate these bias issues. In this survey, we summarise existing methodologies
and practices of fairness in recommender systems. Firstly, we present concepts
of fairness in different recommendation scenarios, comprehensively categorize
current advances, and introduce typical methods to promote fairness in
different stages of recommender systems. Next, after introducing datasets and
evaluation metrics applied to assess the fairness of recommender systems, we
will delve into the significant influence that fairness-aware recommender
systems exert on real-world industrial applications. Subsequently, we highlight
the connection between fairness and other principles of trustworthy recommender
systems, aiming to consider trustworthiness principles holistically while
advocating for fairness. Finally, we summarize this review, spotlighting
promising opportunities in comprehending concepts, frameworks, the balance
between accuracy and fairness, and the ties with trustworthiness, with the
ultimate goal of fostering the development of fairness-aware recommender
systems.Comment: 27 pages, 9 figure
Understanding and Mitigating Multi-sided Exposure Bias in Recommender Systems
Fairness is a critical system-level objective in recommender systems that has
been the subject of extensive recent research. It is especially important in
multi-sided recommendation platforms where it may be crucial to optimize
utilities not just for the end user, but also for other actors such as item
sellers or producers who desire a fair representation of their items. Existing
solutions do not properly address various aspects of multi-sided fairness in
recommendations as they may either solely have one-sided view (i.e. improving
the fairness only for one side), or do not appropriately measure the fairness
for each actor involved in the system. In this thesis, I aim at first
investigating the impact of unfair recommendations on the system and how these
unfair recommendations can negatively affect major actors in the system. Then,
I seek to propose solutions to tackle the unfairness of recommendations. I
propose a rating transformation technique that works as a pre-processing step
before building the recommendation model to alleviate the inherent popularity
bias in the input data and consequently to mitigate the exposure unfairness for
items and suppliers in the recommendation lists. Also, as another solution, I
propose a general graph-based solution that works as a post-processing approach
after recommendation generation for mitigating the multi-sided exposure bias in
the recommendation results. For evaluation, I introduce several metrics for
measuring the exposure fairness for items and suppliers, and show that these
metrics better capture the fairness properties in the recommendation results. I
perform extensive experiments to evaluate the effectiveness of the proposed
solutions. The experiments on different publicly-available datasets and
comparison with various baselines confirm the superiority of the proposed
solutions in improving the exposure fairness for items and suppliers.Comment: Doctoral thesi
Consumer-side Fairness in Recommender Systems: A Systematic Survey of Methods and Evaluation
In the current landscape of ever-increasing levels of digitalization, we are
facing major challenges pertaining to scalability. Recommender systems have
become irreplaceable both for helping users navigate the increasing amounts of
data and, conversely, aiding providers in marketing products to interested
users. The growing awareness of discrimination in machine learning methods has
recently motivated both academia and industry to research how fairness can be
ensured in recommender systems. For recommender systems, such issues are well
exemplified by occupation recommendation, where biases in historical data may
lead to recommender systems relating one gender to lower wages or to the
propagation of stereotypes. In particular, consumer-side fairness, which
focuses on mitigating discrimination experienced by users of recommender
systems, has seen a vast number of diverse approaches for addressing different
types of discrimination. The nature of said discrimination depends on the
setting and the applied fairness interpretation, of which there are many
variations. This survey serves as a systematic overview and discussion of the
current research on consumer-side fairness in recommender systems. To that end,
a novel taxonomy based on high-level fairness interpretation is proposed and
used to categorize the research and their proposed fairness evaluation metrics.
Finally, we highlight some suggestions for the future direction of the field.Comment: Draft submitted to Springer (November 2022
Fairly Adaptive Negative Sampling for Recommendations
Pairwise learning strategies are prevalent for optimizing recommendation
models on implicit feedback data, which usually learns user preference by
discriminating between positive (i.e., clicked by a user) and negative items
(i.e., obtained by negative sampling). However, the size of different item
groups (specified by item attribute) is usually unevenly distributed. We
empirically find that the commonly used uniform negative sampling strategy for
pairwise algorithms (e.g., BPR) can inherit such data bias and oversample the
majority item group as negative instances, severely countering group fairness
on the item side. In this paper, we propose a Fairly adaptive Negative sampling
approach (FairNeg), which improves item group fairness via adaptively adjusting
the group-level negative sampling distribution in the training process. In
particular, it first perceives the model's unfairness status at each step and
then adjusts the group-wise sampling distribution with an adaptive momentum
update strategy for better facilitating fairness optimization. Moreover, a
negative sampling distribution Mixup mechanism is proposed, which gracefully
incorporates existing importance-aware sampling techniques intended for mining
informative negative samples, thus allowing for achieving multiple optimization
purposes. Extensive experiments on four public datasets show our proposed
method's superiority in group fairness enhancement and fairness-utility
tradeoff.Comment: Accepted by TheWebConf202
Multiwinner Voting with Fairness Constraints
Multiwinner voting rules are used to select a small representative subset of
candidates or items from a larger set given the preferences of voters. However,
if candidates have sensitive attributes such as gender or ethnicity (when
selecting a committee), or specified types such as political leaning (when
selecting a subset of news items), an algorithm that chooses a subset by
optimizing a multiwinner voting rule may be unbalanced in its selection -- it
may under or over represent a particular gender or political orientation in the
examples above. We introduce an algorithmic framework for multiwinner voting
problems when there is an additional requirement that the selected subset
should be "fair" with respect to a given set of attributes. Our framework
provides the flexibility to (1) specify fairness with respect to multiple,
non-disjoint attributes (e.g., ethnicity and gender) and (2) specify a score
function. We study the computational complexity of this constrained multiwinner
voting problem for monotone and submodular score functions and present several
approximation algorithms and matching hardness of approximation results for
various attribute group structure and types of score functions. We also present
simulations that suggest that adding fairness constraints may not affect the
scores significantly when compared to the unconstrained case.Comment: The conference version of this paper appears in IJCAI-ECAI 201
In-processing User Constrained Dominant Sets for User-Oriented Fairness in Recommender Systems
Recommender systems are typically biased toward a small group of users,
leading to severe unfairness in recommendation performance, i.e., User-Oriented
Fairness (UOF) issue. The existing research on UOF is limited and fails to deal
with the root cause of the UOF issue: the learning process between advantaged
and disadvantaged users is unfair. To tackle this issue, we propose an
In-processing User Constrained Dominant Sets (In-UCDS) framework, which is a
general framework that can be applied to any backbone recommendation model to
achieve user-oriented fairness. We split In-UCDS into two stages, i.e., the
UCDS modeling stage and the in-processing training stage. In the UCDS modeling
stage, for each disadvantaged user, we extract a constrained dominant set (a
user cluster) containing some advantaged users that are similar to it. In the
in-processing training stage, we move the representations of disadvantaged
users closer to their corresponding cluster by calculating a fairness loss. By
combining the fairness loss with the original backbone model loss, we address
the UOF issue and maintain the overall recommendation performance
simultaneously. Comprehensive experiments on three real-world datasets
demonstrate that In-UCDS outperforms the state-of-the-art methods, leading to a
fairer model with better overall recommendation performance
Fairness in Graph Mining: A Survey
Graph mining algorithms have been playing a significant role in myriad fields
over the years. However, despite their promising performance on various graph
analytical tasks, most of these algorithms lack fairness considerations. As a
consequence, they could lead to discrimination towards certain populations when
exploited in human-centered applications. Recently, algorithmic fairness has
been extensively studied in graph-based applications. In contrast to
algorithmic fairness on independent and identically distributed (i.i.d.) data,
fairness in graph mining has exclusive backgrounds, taxonomies, and fulfilling
techniques. In this survey, we provide a comprehensive and up-to-date
introduction of existing literature under the context of fair graph mining.
Specifically, we propose a novel taxonomy of fairness notions on graphs, which
sheds light on their connections and differences. We further present an
organized summary of existing techniques that promote fairness in graph mining.
Finally, we summarize the widely used datasets in this emerging research field
and provide insights on current research challenges and open questions, aiming
at encouraging cross-breeding ideas and further advances
- …