226 research outputs found
Context-aware Sequential Recommendation
Since sequential information plays an important role in modeling user
behaviors, various sequential recommendation methods have been proposed.
Methods based on Markov assumption are widely-used, but independently combine
several most recent components. Recently, Recurrent Neural Networks (RNN) based
methods have been successfully applied in several sequential modeling tasks.
However, for real-world applications, these methods have difficulty in modeling
the contextual information, which has been proved to be very important for
behavior modeling. In this paper, we propose a novel model, named Context-Aware
Recurrent Neural Networks (CA-RNN). Instead of using the constant input matrix
and transition matrix in conventional RNN models, CA-RNN employs adaptive
context-specific input matrices and adaptive context-specific transition
matrices. The adaptive context-specific input matrices capture external
situations where user behaviors happen, such as time, location, weather and so
on. And the adaptive context-specific transition matrices capture how lengths
of time intervals between adjacent behaviors in historical sequences affect the
transition of global sequential features. Experimental results show that the
proposed CA-RNN model yields significant improvements over state-of-the-art
sequential recommendation methods and context-aware recommendation methods on
two public datasets, i.e., the Taobao dataset and the Movielens-1M dataset.Comment: IEEE International Conference on Data Mining (ICDM) 2016, to apea
Retrospective Higher-Order Markov Processes for User Trails
Users form information trails as they browse the web, checkin with a
geolocation, rate items, or consume media. A common problem is to predict what
a user might do next for the purposes of guidance, recommendation, or
prefetching. First-order and higher-order Markov chains have been widely used
methods to study such sequences of data. First-order Markov chains are easy to
estimate, but lack accuracy when history matters. Higher-order Markov chains,
in contrast, have too many parameters and suffer from overfitting the training
data. Fitting these parameters with regularization and smoothing only offers
mild improvements. In this paper we propose the retrospective higher-order
Markov process (RHOMP) as a low-parameter model for such sequences. This model
is a special case of a higher-order Markov chain where the transitions depend
retrospectively on a single history state instead of an arbitrary combination
of history states. There are two immediate computational advantages: the number
of parameters is linear in the order of the Markov chain and the model can be
fit to large state spaces. Furthermore, by providing a specific structure to
the higher-order chain, RHOMPs improve the model accuracy by efficiently
utilizing history states without risks of overfitting the data. We demonstrate
how to estimate a RHOMP from data and we demonstrate the effectiveness of our
method on various real application datasets spanning geolocation data, review
sequences, and business locations. The RHOMP model uniformly outperforms
higher-order Markov chains, Kneser-Ney regularization, and tensor
factorizations in terms of prediction accuracy
Attentive Neural Architecture Incorporating Song Features For Music Recommendation
Recommender Systems are an integral part of music sharing platforms. Often
the aim of these systems is to increase the time, the user spends on the
platform and hence having a high commercial value. The systems which aim at
increasing the average time a user spends on the platform often need to
recommend songs which the user might want to listen to next at each point in
time. This is different from recommendation systems which try to predict the
item which might be of interest to the user at some point in the user lifetime
but not necessarily in the very near future. Prediction of the next song the
user might like requires some kind of modeling of the user interests at the
given point of time. Attentive neural networks have been exploiting the
sequence in which the items were selected by the user to model the implicit
short-term interests of the user for the task of next item prediction, however
we feel that the features of the songs occurring in the sequence could also
convey some important information about the short-term user interest which only
the items cannot. In this direction, we propose a novel attentive neural
architecture which in addition to the sequence of items selected by the user,
uses the features of these items to better learn the user short-term
preferences and recommend the next song to the user.Comment: Accepted as a paper at the 12th ACM Conference on Recommender Systems
(RecSys 18
- …