57 research outputs found
Embarrassingly Shallow Autoencoders for Sparse Data
Combining simple elements from the literature, we define a linear model that
is geared toward sparse data, in particular implicit feedback data for
recommender systems. We show that its training objective has a closed-form
solution, and discuss the resulting conceptual insights. Surprisingly, this
simple model achieves better ranking accuracy than various state-of-the-art
collaborative-filtering approaches, including deep non-linear models, on most
of the publicly available data-sets used in our experiments.Comment: In the proceedings of the Web Conference (WWW) 2019 (7 pages
On the Dirichlet Prior and Bayesian Regularization
A common objective in learning a model from data is to recover its network structure, while the model parameters are of minor interest. For example, we may wish to recover regulatory networks from high-throughput data sources. In this paper we examine how Bayesian regularization using a Dirichlet prior over the model parameters affects the learned model structure in a domain with discrete variables. Surprisingly, a weak prior in the sense of smaller equivalent sample size leads to a strong regularization of the model structure (sparse graph) given a sufficiently large data set. In particular, the empty graph is obtained in the limit of a vanishing strength of prior belief. This is diametrically opposite to what one may expect in this limit, namely the complete graph from an (unregularized) maximum likelihood estimate. Since the prior affects the parameters as expected, the prior strength balances a "trade-off" between regularizing the parameters or the structure of the model. We demonstrate the benefits of optimizing this trade-off in the sense of predictive accuracy
On Sampling Top-K Recommendation Evaluation
Recently, Rendle has warned that the use of sampling-based top- metrics
might not suffice. This throws a number of recent studies on deep
learning-based recommendation algorithms, and classic non-deep-learning
algorithms using such a metric, into jeopardy. In this work, we thoroughly
investigate the relationship between the sampling and global top- Hit-Ratio
(HR, or Recall), originally proposed by Koren[2] and extensively used by
others. By formulating the problem of aligning sampling top- () and
global top- () Hit-Ratios through a mapping function , so that
, we demonstrate both theoretically and experimentally
that the sampling top- Hit-Ratio provides an accurate approximation of its
global (exact) counterpart, and can consistently predict the correct winners
(the same as indicate by their corresponding global Hit-Ratios)
Large Language Models as Zero-Shot Conversational Recommenders
In this paper, we present empirical studies on conversational recommendation
tasks using representative large language models in a zero-shot setting with
three primary contributions. (1) Data: To gain insights into model behavior in
"in-the-wild" conversational recommendation scenarios, we construct a new
dataset of recommendation-related conversations by scraping a popular
discussion website. This is the largest public real-world conversational
recommendation dataset to date. (2) Evaluation: On the new dataset and two
existing conversational recommendation datasets, we observe that even without
fine-tuning, large language models can outperform existing fine-tuned
conversational recommendation models. (3) Analysis: We propose various probing
tasks to investigate the mechanisms behind the remarkable performance of large
language models in conversational recommendation. We analyze both the large
language models' behaviors and the characteristics of the datasets, providing a
holistic understanding of the models' effectiveness, limitations and suggesting
directions for the design of future conversational recommendersComment: Accepted as CIKM 2023 long paper. Longer version is coming soon
(e.g., more details about dataset
RecVAE: a New Variational Autoencoder for Top-N Recommendations with Implicit Feedback
Recent research has shown the advantages of using autoencoders based on deep
neural networks for collaborative filtering. In particular, the recently
proposed Mult-VAE model, which used the multinomial likelihood variational
autoencoders, has shown excellent results for top-N recommendations. In this
work, we propose the Recommender VAE (RecVAE) model that originates from our
research on regularization techniques for variational autoencoders. RecVAE
introduces several novel ideas to improve Mult-VAE, including a novel composite
prior distribution for the latent codes, a new approach to setting the
hyperparameter for the -VAE framework, and a new approach to training
based on alternating updates. In experimental evaluation, we show that RecVAE
significantly outperforms previously proposed autoencoder-based models,
including Mult-VAE and RaCT, across classical collaborative filtering datasets,
and present a detailed ablation study to assess our new developments. Code and
models are available at https://github.com/ilya-shenbin/RecVAE.Comment: In The Thirteenth ACM International Conference on Web Search and Data
Mining (WSDM '20), February 3-7, 2020, Houston, TX, USA. ACM, New York, NY,
USA, 9 page
- …