57,115 research outputs found
Bias Disparity in Collaborative Recommendation: Algorithmic Evaluation and Comparison
Research on fairness in machine learning has been recently extended to
recommender systems. One of the factors that may impact fairness is bias
disparity, the degree to which a group's preferences on various item categories
fail to be reflected in the recommendations they receive. In some cases biases
in the original data may be amplified or reversed by the underlying
recommendation algorithm. In this paper, we explore how different
recommendation algorithms reflect the tradeoff between ranking quality and bias
disparity. Our experiments include neighborhood-based, model-based, and
trust-aware recommendation algorithms.Comment: Workshop on Recommendation in Multi-Stakeholder Environments (RMSE)
at ACM RecSys 2019, Copenhagen, Denmar
Controlling Fairness and Bias in Dynamic Learning-to-Rank
Rankings are the primary interface through which many online platforms match
users to items (e.g. news, products, music, video). In these two-sided markets,
not only the users draw utility from the rankings, but the rankings also
determine the utility (e.g. exposure, revenue) for the item providers (e.g.
publishers, sellers, artists, studios). It has already been noted that
myopically optimizing utility to the users, as done by virtually all
learning-to-rank algorithms, can be unfair to the item providers. We,
therefore, present a learning-to-rank approach for explicitly enforcing
merit-based fairness guarantees to groups of items (e.g. articles by the same
publisher, tracks by the same artist). In particular, we propose a learning
algorithm that ensures notions of amortized group fairness, while
simultaneously learning the ranking function from implicit feedback data. The
algorithm takes the form of a controller that integrates unbiased estimators
for both fairness and utility, dynamically adapting both as more data becomes
available. In addition to its rigorous theoretical foundation and convergence
guarantees, we find empirically that the algorithm is highly practical and
robust.Comment: First two authors contributed equally. In Proceedings of the 43rd
International ACM SIGIR Conference on Research and Development in Information
Retrieval 202
Human Relations Report
Any assessment of the state of human relations in the Chicago region needs to be multidimensional. At its most basic, such an assessment involves the quality of relationships, or relations, among individuals. Relations may manifest themselves in families, among friends, within neighborhoods, or in work, religious, educational, recreational or other social settings. There are no widely accepted measures of the quality of human relations, in part because different commentators view the subject differently. Quality human relations may have several outcomes: for people to be satisfied or experience a good quality of life; for people to be supportive and helpful to one another; or for people to treat one another fairly and equally.In some social settings, individuals with common characteristics share a common fate or have similar life experiences and opportunities. Other social settings are marked more by differences among groups than commonalities. Such differences can be readily observed in the cases of different racial, ethnic, age or language groups; among persons sharing a gender or sexual orientation; or among the disabled. These social groupings seem to have the most impact on people's condition and identity
On bias in social reviews of university courses
University course ranking forums are a popular means of disseminating
information about satisfaction with the quality of course content and
instruction, especially with undergraduate students. A variety of policy
decisions by university administrators, instructional designers and teaching
staff affect how students perceive the efficacy of pedagogies employed in a
given course, in class and online. While there is a large body of research on
qualitative driving factors behind the use of academic rating sites, there is
little investigation of the (potential) implicit student bias on said forums
towards desirable course outcomes at the institution level. To that end, we
examine the connection between course outcomes (student-reported GPA) and the
overall ranking of the primary course instructor, as well as rating disparity
by nature of course outcomes, for several hundred courses taught at Virginia
Tech based on data collected from a popular academic rating forum. We also
replicate our analysis for several public universities across the US. Our
experiments indicate that there is a discernible albeit complex bias towards
course outcomes in the professor ratings registered by students.Comment: WebSci'19 Companion Proceeding
Modeling and debiasing feedback loops in collaborative filtering recommender systems.
Artificial Intelligence (AI)-driven recommender systems have been gaining increasing ubiquity and influence in our daily lives, especially during time spent online on the World Wide Web or smart devices. The influence of recommender systems on who and what we can find and discover, our choices, and our behavior, has thus never been more concrete. AI can now predict and anticipate, with varying degrees of accuracy, the news article we will read, the music we will listen to, the movies we will watch, the transactions we will make, the restaurants we will eat in, the online courses we will be interested in, and the people we will connect with for various ends and purposes. For all these reasons, the automated predictions and recommendations made by AI can lead to influencing and changing human opinions, behavior, and decision making. When the AI predictions are biased, the influences can have unfair consequences on society, ranging from social polarization to the amplification of misinformation and hate speech. For instance, bias in recommender systems can affect the decision making and shift consumer behavior in an unfair way due to a phenomenon known as the feedback loop. The feedback loop is an inherent component of recommender systems because the latter are dynamic systems that involve continuous interactions with the users, whereby data collected to train a recommender system model is usually affected by the outputs of a previously trained model. This feedback loop is expected to affect the performance of the system. For instance, it can amplify initial bias in the data or model and can lead to other phenomena such as filter bubbles, polarization, and popularity bias. Up to now, it has been difficult to understand the dynamics of recommender system feedback loops, and equally challenging to evaluate the bias and filter bubbles emerging from recommender system models within such an iterative closed loop environment. In this dissertation, we study the feedback loop in the context of Collaborative Filtering (CF) recommender systems. CF systems comprise the leading family of recommender systems that rely mainly on mining the patterns of interaction between the users and items to train models that aim to predict future user interactions. Our research contributions target three aspects of recommendation, namely modeling, debiasing and evaluating feedback loops. Our research advances the state of the art in Fairness in Artificial Intelligence on several fronts: (1) We propose and validate a new theoretical model, based on Martingale differences, to model the recommender system feedback loop, and allow a better understanding of the dynamics of filter bubbles and user discovery. (2) We propose a Transformer-based deep learning architecture and algorithm to learn diverse representations for users and items in order to increase the diversity in the recommendations. Our evaluation experiments on real world datasets demonstrate that our transformer model recommends 14\% more diverse items and improves the novelty of the recommendation by more than 20\%. (3) We propose a new simulation and experimentation framework that allows studying and tracking the evolution of bias metrics in a feedback loop setting, for a variety of recommendation modeling algorithms. Our preliminary findings, using the new simulation framework show that recommender systems are deeply affected by the feedback loop, and that without an adequate debiasing or exploration strategy, this feedback loop limits the discovery of the user and increases the disparity in exposure between items that can be recommended. To help the research and practice community in studying recommender system fairness, all the tools developed to model, debias, and evaluate recommender systems are made available to the public as open source software libraries \footnote{https://github.com/samikhenissi/TheoretUserModeling}. (4) We propose a novel learnable dynamic debiasing strategy that learns an optimal rescaling parameter for the predicted rating and achieves a better trade-off between accuracy and debiasing. We focus on solving the popularity bias of the items and test our method using our proposed simulation framework and show the effectiveness of using a learnable debiasing degree to produce better results
- …