6,453 research outputs found
Towards Explainable Conversational Recommender Systems
Explanations in conventional recommender systems have demonstrated benefits
in helping the user understand the rationality of the recommendations and
improving the system's efficiency, transparency, and trustworthiness. In the
conversational environment, multiple contextualized explanations need to be
generated, which poses further challenges for explanations. To better measure
explainability in conversational recommender systems (CRS), we propose ten
evaluation perspectives based on concepts from conventional recommender systems
together with the characteristics of CRS. We assess five existing CRS benchmark
datasets using these metrics and observe the necessity of improving the
explanation quality of CRS. To achieve this, we conduct manual and automatic
approaches to extend these dialogues and construct a new CRS dataset, namely
Explainable Recommendation Dialogues (E-ReDial). It includes 756 dialogues with
over 2,000 high-quality rewritten explanations. We compare two baseline
approaches to perform explanation generation based on E-ReDial. Experimental
results suggest that models trained on E-ReDial can significantly improve
explainability while introducing knowledge into the models can further improve
the performance. GPT-3 in the in-context learning setting can generate more
realistic and diverse movie descriptions. In contrast, T5 training on E-ReDial
can better generate clear reasons for recommendations based on user
preferences. E-ReDial is available at https://github.com/Superbooming/E-ReDial
Mining semantic knowledge graphs to add explainability to black box recommender systems
Recommender systems are being increasingly used to predict the preferences of users on online platforms and recommend relevant options that help them cope with information overload. In particular, modern model-based collaborative filtering algorithms, such as latent factor models, are considered state-of-the-art in recommendation systems. Unfortunately, these black box systems lack transparency, as they provide little information about the reasoning behind their predictions. White box systems, in contrast, can, by nature, easily generate explanations. However, their predictions are less accurate than sophisticated black box models. Recent research has demonstrated that explanations are an essential component in bringing the powerful predictions of big data and machine learning methods to a mass audience without compromising trust. Explanations can take a variety of formats, depending on the recommendation domain and the machine learning model used to make predictions. The objective of this work is to build a recommender system that can generate both accurate predictions and semantically rich explanations that justify the predictions. We propose a novel approach to build an explanation generation mechanism into a latent factor-based black box recommendation model. The designed model is trained to learn to make predictions that are accompanied by explanations that are automatically mined from the semantic web. Our evaluation experiments, which carefully study the trade-offs between the quality of predictions and explanations, show that our proposed approach succeeds in producing explainable predictions without a significant sacrifice in prediction accuracy
SeER: An Explainable Deep Learning MIDI-based Hybrid Song Recommender System
State of the art music recommender systems mainly rely on either matrix factorization-based collaborative filtering approaches or deep learning architectures. Deep learning models usually use metadata for content-based filtering or predict the next user interaction by learning from temporal sequences of user actions. Despite advances in deep learning for song recommendation, none has taken advantage of the sequential nature of songs by learning sequence models that are based on content. Aside from the importance of prediction accuracy, other significant aspects are important, such as explainability and solving the cold start problem. In this work, we propose a hybrid deep learning model, called “SeER , that uses collaborative filtering (CF) and deep learning sequence models on the MIDI content of songs for recommendation in order to provide more accurate personalized recommendations; solve the item cold start problem; and generate a relevant explanation for a song recommendation. Our evaluation experiments show promising results compared to state of the art baseline and hybrid song recommender systems in terms of ranking evaluation. Moreover, based on proposed tests for offline validation, we show that our personalized explanations capture properties that are in accordance with the user’s preferences
Scalable and interpretable product recommendations via overlapping co-clustering
We consider the problem of generating interpretable recommendations by
identifying overlapping co-clusters of clients and products, based only on
positive or implicit feedback. Our approach is applicable on very large
datasets because it exhibits almost linear complexity in the input examples and
the number of co-clusters. We show, both on real industrial data and on
publicly available datasets, that the recommendation accuracy of our algorithm
is competitive to that of state-of-art matrix factorization techniques. In
addition, our technique has the advantage of offering recommendations that are
textually and visually interpretable. Finally, we examine how to implement our
technique efficiently on Graphical Processing Units (GPUs).Comment: In IEEE International Conference on Data Engineering (ICDE) 201
- …