673 research outputs found
New debiasing strategies in collaborative filtering recommender systems: modeling user conformity, multiple biases, and causality.
Recommender Systems are widely used to personalize the user experience in a diverse set of online applications ranging from e-commerce and education to social media and online entertainment. These State of the Art AI systems can suffer from several biases that may occur at different stages of the recommendation life-cycle. For instance, using biased data to train recommendation models may lead to several issues, such as the discrepancy between online and offline evaluation, decreasing the recommendation performance, and hurting the user experience. Bias can occur during the data collection stage where the data inherits the user-item interaction biases, such as selection and exposure bias. Bias can also occur in the training stage, where popular items tend to be recommended much more frequently given that they received more interactions to start with. The closed feedback loop nature of online recommender systems will further amplify the latter biases as well. In this dissertation, we study the bias in the context of Collaborative Filtering recommender system, and propose a new Popularity Correction Matrix Factorization (PCMF) that aims to improve the recommender system performance as well as decrease popularity bias and increase the diversity of items in the recommendation lists. PCMF mitigates popularity bias by disentangling relevance and conformity and by learning a user-personalized bias vector to capture the users\u27 individual conformity levels along a full spectrum of conformity bias. One shortcoming of the proposed PCMF debiasing approach, is its assumption that the recommender system is affected by only popularity bias. However in the real word, different types of bias do occur simultaneously and interact with one another. We therefore relax the latter assumption and propose a multi-pronged approach that can account for two biases simultaneously, namely popularity and exposure bias. our experimental results show that accounting for multiple biases does improve the results in terms of providing more accurate and less biased results. Finally, we propose a novel two-stage debiasing approach, inspired from the proximal causal inference framework. Unlike the existing causal IPS approach that corrects for observed confounders, our proposed approach corrects for both observed and potential unobserved confounders. The approach relies on a pair of negative control variables to adjust for the bias in the potential ratings. Our proposed approach outperforms state of the art causal approaches, proving that accounting for unobserved confounders can improve the recommendation system\u27s performance
Mitigation of Popularity Bias in Recommendation Systems
In response to the quantity of information available on the Internet, many online service providers are attempting to customize their services and make content access more simple via recommender systems (RSs) to support users in discovering the products they are most likely interested in. However, these recommendation systems are prone to popularity bias, which is a tendency to promote popular items even if they do not satisfy a user’s preferences and then provide customers with recommendations of poor quality. Such a bias has a negative influence on both users and item providers. It is then essential to mitigate such bias in order to guarantee that less popular but pertinent items show up on the user’s
recommendation list. In this work, we conduct an empirical analysis of different mitigation techniques for popularity bias to provide an overview of the present state of the art of popularity bias and raise the fairness issue in RSs
Causal Disentanglement for Semantics-Aware Intent Learning in Recommendation
Traditional recommendation models trained on observational interaction data have generated large impacts in a wide range of applications, it faces bias problems that cover users true intent and thus deteriorate the recommendation effectiveness. Existing methods tracks this problem as eliminating bias for the robust recommendation, e.g., by re-weighting training samples or learning disentangled representation. The disentangled representation methods as the state-of-the-art eliminate bias through revealing cause-effect of the bias generation. However, how to design the semantics-aware and unbiased representation for users true intents is largely unexplored. To bridge the gap, we are the first to propose an unbiased and semantics-aware disentanglement learning called CaDSI (Causal Disentanglement for Semantics-Aware Intent Learning) from a causal perspective. Particularly, CaDSI explicitly models the causal relations underlying recommendation task, and thus produces semantics-aware representations via disentangling users true intents aware of specific item context. Moreover, the causal intervention mechanism is designed to eliminate confounding bias stemmed from context information, which further to align the semantics-aware representation with users true intent. Extensive experiments and case studies both validate the robustness and interpretability of our proposed model
CAM2: Conformity-Aware Multi-Task Ranking Model for Large-Scale Recommender Systems
Learning large-scale industrial recommender system models by fitting them to
historical user interaction data makes them vulnerable to conformity bias. This
may be due to a number of factors, including the fact that user interests may
be difficult to determine and that many items are often interacted with based
on ecosystem factors other than their relevance to the individual user. In this
work, we introduce CAM2, a conformity-aware multi-task ranking model to serve
relevant items to users on one of the largest industrial recommendation
platforms. CAM2 addresses these challenges systematically by leveraging causal
modeling to disentangle users' conformity to popular items from their true
interests. This framework is generalizable and can be scaled to support
multiple representations of conformity and user relevance in any large-scale
recommender system. We provide deeper practical insights and demonstrate the
effectiveness of the proposed model through improvements in offline evaluation
metrics compared to our production multi-task ranking model. We also show
through online experiments that the CAM2 model results in a significant 0.50%
increase in aggregated user engagement, coupled with a 0.21% increase in daily
active users on Facebook Watch, a popular video discovery and sharing platform
serving billions of users.Comment: Accepted by WWW 202
Capturing Popularity Trends: A Simplistic Non-Personalized Approach for Enhanced Item Recommendation
Recommender systems have been gaining increasing research attention over the
years. Most existing recommendation methods focus on capturing users'
personalized preferences through historical user-item interactions, which may
potentially violate user privacy. Additionally, these approaches often overlook
the significance of the temporal fluctuation in item popularity that can sway
users' decision-making. To bridge this gap, we propose Popularity-Aware
Recommender (PARE), which makes non-personalized recommendations by predicting
the items that will attain the highest popularity. PARE consists of four
modules, each focusing on a different aspect: popularity history, temporal
impact, periodic impact, and side information. Finally, an attention layer is
leveraged to fuse the outputs of four modules. To our knowledge, this is the
first work to explicitly model item popularity in recommendation systems.
Extensive experiments show that PARE performs on par or even better than
sophisticated state-of-the-art recommendation methods. Since PARE prioritizes
item popularity over personalized user preferences, it can enhance existing
recommendation methods as a complementary component. Our experiments
demonstrate that integrating PARE with existing recommendation methods
significantly surpasses the performance of standalone models, highlighting
PARE's potential as a complement to existing recommendation methods.
Furthermore, the simplicity of PARE makes it immensely practical for industrial
applications and a valuable baseline for future research.Comment: 9 pages, 5 figure
Mitigating Popularity Bias in Recommendation with Unbalanced Interactions: A Gradient Perspective
Recommender systems learn from historical user-item interactions to identify
preferred items for target users. These observed interactions are usually
unbalanced following a long-tailed distribution. Such long-tailed data lead to
popularity bias to recommend popular but not personalized items to users. We
present a gradient perspective to understand two negative impacts of popularity
bias in recommendation model optimization: (i) the gradient direction of
popular item embeddings is closer to that of positive interactions, and (ii)
the magnitude of positive gradient for popular items are much greater than that
of unpopular items. To address these issues, we propose a simple yet efficient
framework to mitigate popularity bias from a gradient perspective.
Specifically, we first normalize each user embedding and record accumulated
gradients of users and items via popularity bias measures in model training. To
address the popularity bias issues, we develop a gradient-based embedding
adjustment approach used in model testing. This strategy is generic,
model-agnostic, and can be seamlessly integrated into most existing recommender
systems. Our extensive experiments on two classic recommendation models and
four real-world datasets demonstrate the effectiveness of our method over
state-of-the-art debiasing baselines.Comment: Recommendation System, Popularity Bia
Alleviating Video-Length Effect for Micro-video Recommendation
Micro-videos platforms such as TikTok are extremely popular nowadays. One
important feature is that users no longer select interested videos from a set,
instead they either watch the recommended video or skip to the next one. As a
result, the time length of users' watching behavior becomes the most important
signal for identifying preferences. However, our empirical data analysis has
shown a video-length effect that long videos are easier to receive a higher
value of average view time, thus adopting such view-time labels for measuring
user preferences can easily induce a biased model that favors the longer
videos. In this paper, we propose a Video Length Debiasing Recommendation
(VLDRec) method to alleviate such an effect for micro-video recommendation.
VLDRec designs the data labeling approach and the sample generation module that
better capture user preferences in a view-time oriented manner. It further
leverages the multi-task learning technique to jointly optimize the above
samples with original biased ones. Extensive experiments show that VLDRec can
improve the users' view time by 1.81% and 11.32% on two real-world datasets,
given a recommendation list of a fixed overall video length, compared with the
best baseline method. Moreover, VLDRec is also more effective in matching
users' interests in terms of the video content.Comment: Accept by TOI
Reducing Popularity Bias in Recommender Systems through AUC-Optimal Negative Sampling
Popularity bias is a persistent issue associated with recommendation systems,
posing challenges to both fairness and efficiency. Existing literature widely
acknowledges that reducing popularity bias often requires sacrificing
recommendation accuracy. In this paper, we challenge this commonly held belief.
Our analysis under general bias-variance decomposition framework shows that
reducing bias can actually lead to improved model performance under certain
conditions. To achieve this win-win situation, we propose to intervene in model
training through negative sampling thereby modifying model predictions.
Specifically, we provide an optimal negative sampling rule that maximizes
partial AUC to preserve the accuracy of any given model, while correcting
sample information and prior information to reduce popularity bias in a
flexible and principled way. Our experimental results on real-world datasets
demonstrate the superiority of our approach in improving recommendation
performance and reducing popularity bias.Comment: 20 page
Evaluation Framework for Understanding Sensitive Attribute Association Bias in Latent Factor Recommendation Algorithms
We present a novel evaluation framework for representation bias in latent
factor recommendation (LFR) algorithms. Our framework introduces the concept of
attribute association bias in recommendations allowing practitioners to explore
how recommendation systems can introduce or amplify stakeholder representation
harm. Attribute association bias (AAB) occurs when sensitive attributes become
semantically captured or entangled in the trained recommendation latent space.
This bias can result in the recommender reinforcing harmful stereotypes, which
may result in downstream representation harms to system consumer and provider
stakeholders. LFR models are at risk of experiencing AAB due to their ability
to entangle explicit and implicit attributes into the trained latent space.
Understanding this phenomenon is essential due to the increasingly common use
of entity vectors as attributes in downstream components in hybrid industry
recommendation systems. We provide practitioners with a framework for executing
disaggregated evaluations of AAB within broader algorithmic auditing
frameworks. Inspired by research in natural language processing (NLP) observing
gender bias in word embeddings, our framework introduces AAB evaluation methods
specifically for recommendation entity vectors. We present four evaluation
strategies for sensitive AAB in LFR models: attribute bias directions,
attribute association bias metrics, classification for explaining bias, and
latent space visualization. We demonstrate the utility of our framework by
evaluating user gender AAB regarding podcast genres with an industry case study
of a production-level DNN recommendation model. We uncover significant levels
of user gender AAB when user gender is used and removed as a model feature
during training, pointing to the potential for systematic bias in LFR model
outputs
- …