1,309 research outputs found
Adaptive Vague Preference Policy Learning for Multi-round Conversational Recommendation
Conversational recommendation systems (CRS) effectively address information
asymmetry by dynamically eliciting user preferences through multi-turn
interactions. Existing CRS widely assumes that users have clear preferences.
Under this assumption, the agent will completely trust the user feedback and
treat the accepted or rejected signals as strong indicators to filter items and
reduce the candidate space, which may lead to the problem of over-filtering.
However, in reality, users' preferences are often vague and volatile, with
uncertainty about their desires and changing decisions during interactions.
To address this issue, we introduce a novel scenario called Vague Preference
Multi-round Conversational Recommendation (VPMCR), which considers users' vague
and volatile preferences in CRS.VPMCR employs a soft estimation mechanism to
assign a non-zero confidence score for all candidate items to be displayed,
naturally avoiding the over-filtering problem. In the VPMCR setting, we
introduce an solution called Adaptive Vague Preference Policy Learning (AVPPL),
which consists of two main components: Uncertainty-aware Soft Estimation (USE)
and Uncertainty-aware Policy Learning (UPL). USE estimates the uncertainty of
users' vague feedback and captures their dynamic preferences using a
choice-based preferences extraction module and a time-aware decaying strategy.
UPL leverages the preference distribution estimated by USE to guide the
conversation and adapt to changes in users' preferences to make recommendations
or ask for attributes.
Our extensive experiments demonstrate the effectiveness of our method in the
VPMCR scenario, highlighting its potential for practical applications and
improving the overall performance and applicability of CRS in real-world
settings, particularly for users with vague or dynamic preferences
Adapting Progress Feedback and Emotional Support to Learner Personality
Peer reviewedPostprin
Meta Policy Learning for Cold-Start Conversational Recommendation
Conversational recommender systems (CRS) explicitly solicit users'
preferences for improved recommendations on the fly. Most existing CRS
solutions count on a single policy trained by reinforcement learning for a
population of users. However, for users new to the system, such a global policy
becomes ineffective to satisfy them, i.e., the cold-start challenge. In this
paper, we study CRS policy learning for cold-start users via meta-reinforcement
learning. We propose to learn a meta policy and adapt it to new users with only
a few trials of conversational recommendations. To facilitate fast policy
adaptation, we design three synergetic components. Firstly, we design a
meta-exploration policy dedicated to identifying user preferences via a few
exploratory conversations, which accelerates personalized policy adaptation
from the meta policy. Secondly, we adapt the item recommendation module for
each user to maximize the recommendation quality based on the collected
conversation states during conversations. Thirdly, we propose a
Transformer-based state encoder as the backbone to connect the previous two
components. It provides comprehensive state representations by modeling
complicated relations between positive and negative feedback during the
conversation. Extensive experiments on three datasets demonstrate the advantage
of our solution in serving new users, compared with a rich set of
state-of-the-art CRS solutions.Comment: 10 pages, WSDM202
How to Retrain Recommender System? A Sequential Meta-Learning Method
Practical recommender systems need be periodically retrained to refresh the
model with new interaction data. To pursue high model fidelity, it is usually
desirable to retrain the model on both historical and new data, since it can
account for both long-term and short-term user preference. However, a full
model retraining could be very time-consuming and memory-costly, especially
when the scale of historical data is large. In this work, we study the model
retraining mechanism for recommender systems, a topic of high practical values
but has been relatively little explored in the research community.
Our first belief is that retraining the model on historical data is
unnecessary, since the model has been trained on it before. Nevertheless,
normal training on new data only may easily cause overfitting and forgetting
issues, since the new data is of a smaller scale and contains fewer information
on long-term user preference. To address this dilemma, we propose a new
training method, aiming to abandon the historical data during retraining
through learning to transfer the past training experience. Specifically, we
design a neural network-based transfer component, which transforms the old
model to a new model that is tailored for future recommendations. To learn the
transfer component well, we optimize the "future performance" -- i.e., the
recommendation accuracy evaluated in the next time period. Our Sequential
Meta-Learning(SML) method offers a general training paradigm that is applicable
to any differentiable model. We demonstrate SML on matrix factorization and
conduct experiments on two real-world datasets. Empirical results show that SML
not only achieves significant speed-up, but also outperforms the full model
retraining in recommendation accuracy, validating the effectiveness of our
proposals. We release our codes at: https://github.com/zyang1580/SML.Comment: Appear in SIGIR 202
Tidying Up the Conversational Recommender Systems' Biases
The growing popularity of language models has sparked interest in
conversational recommender systems (CRS) within both industry and research
circles. However, concerns regarding biases in these systems have emerged.
While individual components of CRS have been subject to bias studies, a
literature gap remains in understanding specific biases unique to CRS and how
these biases may be amplified or reduced when integrated into complex CRS
models. In this paper, we provide a concise review of biases in CRS by
surveying recent literature. We examine the presence of biases throughout the
system's pipeline and consider the challenges that arise from combining
multiple models. Our study investigates biases in classic recommender systems
and their relevance to CRS. Moreover, we address specific biases in CRS,
considering variations with and without natural language understanding
capabilities, along with biases related to dialogue systems and language
models. Through our findings, we highlight the necessity of adopting a holistic
perspective when dealing with biases in complex CRS models
A Conversation is Worth A Thousand Recommendations: A Survey of Holistic Conversational Recommender Systems
Conversational recommender systems (CRS) generate recommendations through an
interactive process. However, not all CRS approaches use human conversations as
their source of interaction data; the majority of prior CRS work simulates
interactions by exchanging entity-level information. As a result, claims of
prior CRS work do not generalise to real-world settings where conversations
take unexpected turns, or where conversational and intent understanding is not
perfect. To tackle this challenge, the research community has started to
examine holistic CRS, which are trained using conversational data collected
from real-world scenarios. Despite their emergence, such holistic approaches
are under-explored.
We present a comprehensive survey of holistic CRS methods by summarizing the
literature in a structured manner. Our survey recognises holistic CRS
approaches as having three components: 1) a backbone language model, the
optional use of 2) external knowledge, and/or 3) external guidance. We also
give a detailed analysis of CRS datasets and evaluation methods in real
application scenarios. We offer our insight as to the current challenges of
holistic CRS and possible future trends.Comment: Accepted by 5th KaRS Workshop @ ACM RecSys 2023, 8 page
Adapting Feedback to Personality to Increase Motivation
Peer reviewedPostprin
- ā¦