3,371 research outputs found
Top-N Recommendation on Graphs
Recommender systems play an increasingly important role in online
applications to help users find what they need or prefer. Collaborative
filtering algorithms that generate predictions by analyzing the user-item
rating matrix perform poorly when the matrix is sparse. To alleviate this
problem, this paper proposes a simple recommendation algorithm that fully
exploits the similarity information among users and items and intrinsic
structural information of the user-item matrix. The proposed method constructs
a new representation which preserves affinity and structure information in the
user-item rating matrix and then performs recommendation task. To capture
proximity information about users and items, two graphs are constructed.
Manifold learning idea is used to constrain the new representation to be smooth
on these graphs, so as to enforce users and item proximities. Our model is
formulated as a convex optimization problem, for which we need to solve the
well-known Sylvester equation only. We carry out extensive empirical
evaluations on six benchmark datasets to show the effectiveness of this
approach.Comment: CIKM 201
A Collective Variational Autoencoder for Top- Recommendation with Side Information
Recommender systems have been studied extensively due to their practical use
in many real-world scenarios. Despite this, generating effective
recommendations with sparse user ratings remains a challenge. Side information
associated with items has been widely utilized to address rating sparsity.
Existing recommendation models that use side information are linear and, hence,
have restricted expressiveness. Deep learning has been used to capture
non-linearities by learning deep item representations from side information but
as side information is high-dimensional existing deep models tend to have large
input dimensionality, which dominates their overall size. This makes them
difficult to train, especially with small numbers of inputs.
Rather than learning item representations, which is problematic with
high-dimensional side information, in this paper, we propose to learn feature
representation through deep learning from side information. Learning feature
representations, on the other hand, ensures a sufficient number of inputs to
train a deep network. To achieve this, we propose to simultaneously recover
user ratings and side information, by using a Variational Autoencoder (VAE).
Specifically, user ratings and side information are encoded and decoded
collectively through the same inference network and generation network. This is
possible as both user ratings and side information are data associated with
items. To account for the heterogeneity of user rating and side information,
the final layer of the generation network follows different distributions
depending on the type of information. The proposed model is easy to implement
and efficient to optimize and is shown to outperform state-of-the-art top-
recommendation methods that use side information.Comment: 7 pages, 3 figures, DLRS workshop 201
Privacy-preserving distributed service recommendation based on locality-sensitive hashing
With the advent of IoT (Internet of Things) age, considerable web services are emerging rapidly in service communities, which places a heavy burden on the target users’ service selection decisions. In this situation, various techniques, e.g., collaborative filtering (i.e., CF) is introduced in service recommendation to alleviate the service selection burden. However, traditional CF-based service recommendation approaches often assume that the historical user-service quality data is centralized, while neglect the distributed recommendation situation. Generally, distributed service recommendation involves inevitable message communication among different parties and hence, brings challenging efficiency and privacy concerns. In view of this challenge, a novel privacy-preserving distributed service recommendation approach based on Locality-Sensitive Hashing (LSH), i.e., DistSRLSH is put forward in this paper. Through LSH, DistSRLSH can achieve a good tradeoff among service recommendation accuracy, privacy-preservation and efficiency in distributed environment. Finally, through a set of experiments deployed on WS-DREAM dataset, we validate the feasibility of our proposal in handling distributed service recommendation problems
Learning compact hashing codes with complex objectives from multiple sources for large scale similarity search
Similarity search is a key problem in many real world applications including image and text retrieval, content reuse detection and collaborative filtering. The purpose of similarity search is to identify similar data examples given a query example. Due to the explosive growth of the Internet, a huge amount of data such as texts, images and videos has been generated, which indicates that efficient large scale similarity search becomes more important.^ Hashing methods have become popular for large scale similarity search due to their computational and memory efficiency. These hashing methods design compact binary codes to represent data examples so that similar examples are mapped into similar codes. This dissertation addresses five major problems for utilizing supervised information from multiple sources in hashing with respect to different objectives. Firstly, we address the problem of incorporating semantic tags by modeling the latent correlations between tags and data examples. More precisely, the hashing codes are learned in a unified semi-supervised framework by simultaneously preserving the similarities between data examples and ensuring the tag consistency via a latent factor model. Secondly, we solve the missing data problem by latent subspace learning from multiple sources. The hashing codes are learned by enforcing the data consistency among different sources. Thirdly, we address the problem of hashing on structured data by graph learning. A weighted graph is constructed based on the structured knowledge from the data. The hashing codes are then learned by preserving the graph similarities. Fourthly, we address the problem of learning high ranking quality hashing codes by utilizing the relevance judgments from users. The hashing code/function is learned via optimizing a commonly used non-smooth non-convex ranking measure, NDCG. Finally, we deal with the problem of insufficient supervision by active learning. We propose to actively select the most informative data examples and tags in a joint manner based on the selection criteria that both the data examples and tags should be most uncertain and dissimilar with each other.^ Extensive experiments on several large scale datasets demonstrate the superior performance of the proposed approaches over several state-of-the-art hashing methods from different perspectives
Discrete Factorization Machines for Fast Feature-based Recommendation
User and item features of side information are crucial for accurate
recommendation. However, the large number of feature dimensions, e.g., usually
larger than 10^7, results in expensive storage and computational cost. This
prohibits fast recommendation especially on mobile applications where the
computational resource is very limited. In this paper, we develop a generic
feature-based recommendation model, called Discrete Factorization Machine
(DFM), for fast and accurate recommendation. DFM binarizes the real-valued
model parameters (e.g., float32) of every feature embedding into binary codes
(e.g., boolean), and thus supports efficient storage and fast user-item score
computation. To avoid the severe quantization loss of the binarization, we
propose a convergent updating rule that resolves the challenging discrete
optimization of DFM. Through extensive experiments on two real-world datasets,
we show that 1) DFM consistently outperforms state-of-the-art binarized
recommendation models, and 2) DFM shows very competitive performance compared
to its real-valued version (FM), demonstrating the minimized quantization loss.
This work is accepted by IJCAI 2018.Comment: Appeared in IJCAI 201
- …