596 research outputs found
Latent Relational Metric Learning via Memory-based Attention for Collaborative Ranking
This paper proposes a new neural architecture for collaborative ranking with
implicit feedback. Our model, LRML (\textit{Latent Relational Metric Learning})
is a novel metric learning approach for recommendation. More specifically,
instead of simple push-pull mechanisms between user and item pairs, we propose
to learn latent relations that describe each user item interaction. This helps
to alleviate the potential geometric inflexibility of existing metric learing
approaches. This enables not only better performance but also a greater extent
of modeling capability, allowing our model to scale to a larger number of
interactions. In order to do so, we employ a augmented memory module and learn
to attend over these memory blocks to construct latent relations. The
memory-based attention module is controlled by the user-item interaction,
making the learned relation vector specific to each user-item pair. Hence, this
can be interpreted as learning an exclusive and optimal relational translation
for each user-item interaction. The proposed architecture demonstrates the
state-of-the-art performance across multiple recommendation benchmarks. LRML
outperforms other metric learning models by in terms of Hits@10 and
nDCG@10 on large datasets such as Netflix and MovieLens20M. Moreover,
qualitative studies also demonstrate evidence that our proposed model is able
to infer and encode explicit sentiment, temporal and attribute information
despite being only trained on implicit feedback. As such, this ascertains the
ability of LRML to uncover hidden relational structure within implicit
datasets.Comment: WWW 201
CoNet: Collaborative Cross Networks for Cross-Domain Recommendation
The cross-domain recommendation technique is an effective way of alleviating
the data sparse issue in recommender systems by leveraging the knowledge from
relevant domains. Transfer learning is a class of algorithms underlying these
techniques. In this paper, we propose a novel transfer learning approach for
cross-domain recommendation by using neural networks as the base model. In
contrast to the matrix factorization based cross-domain techniques, our method
is deep transfer learning, which can learn complex user-item interaction
relationships. We assume that hidden layers in two base networks are connected
by cross mappings, leading to the collaborative cross networks (CoNet). CoNet
enables dual knowledge transfer across domains by introducing cross connections
from one base network to another and vice versa. CoNet is achieved in
multi-layer feedforward networks by adding dual connections and joint loss
functions, which can be trained efficiently by back-propagation. The proposed
model is thoroughly evaluated on two large real-world datasets. It outperforms
baselines by relative improvements of 7.84\% in NDCG. We demonstrate the
necessity of adaptively selecting representations to transfer. Our model can
reduce tens of thousands training examples comparing with non-transfer methods
and still has the competitive performance with them.Comment: Deep transfer learning for recommender system
New Approaches in Multi-View Clustering
Many real-world datasets can be naturally described by multiple views. Due to this, multi-view learning has drawn much attention from both academia and industry. Compared to single-view learning, multi-view learning has demonstrated plenty of advantages. Clustering has long been serving as a critical technique in data mining and machine learning. Recently, multi-view clustering has achieved great success in various applications. To provide a comprehensive review of the typical multi-view clustering methods and their corresponding recent developments, this chapter summarizes five kinds of popular clustering methods and their multi-view learning versions, which include k-means, spectral clustering, matrix factorization, tensor decomposition, and deep learning. These clustering methods are the most widely employed algorithms for single-view data, and lots of efforts have been devoted to extending them for multi-view clustering. Besides, many other multi-view clustering methods can be unified into the frameworks of these five methods. To promote further research and development of multi-view clustering, some popular and open datasets are summarized in two categories. Furthermore, several open issues that deserve more exploration are pointed out in the end
Self-Adaptive Hierarchical Sentence Model
The ability to accurately model a sentence at varying stages (e.g.,
word-phrase-sentence) plays a central role in natural language processing. As
an effort towards this goal we propose a self-adaptive hierarchical sentence
model (AdaSent). AdaSent effectively forms a hierarchy of representations from
words to phrases and then to sentences through recursive gated local
composition of adjacent segments. We design a competitive mechanism (through
gating networks) to allow the representations of the same sentence to be
engaged in a particular learning task (e.g., classification), therefore
effectively mitigating the gradient vanishing problem persistent in other
recursive models. Both qualitative and quantitative analysis shows that AdaSent
can automatically form and select the representations suitable for the task at
hand during training, yielding superior classification performance over
competitor models on 5 benchmark data sets.Comment: 8 pages, 7 figures, accepted as a full paper at IJCAI 201
Multi-view Fuzzy Representation Learning with Rules based Model
Unsupervised multi-view representation learning has been extensively studied
for mining multi-view data. However, some critical challenges remain. On the
one hand, the existing methods cannot explore multi-view data comprehensively
since they usually learn a common representation between views, given that
multi-view data contains both the common information between views and the
specific information within each view. On the other hand, to mine the nonlinear
relationship between data, kernel or neural network methods are commonly used
for multi-view representation learning. However, these methods are lacking in
interpretability. To this end, this paper proposes a new multi-view fuzzy
representation learning method based on the interpretable Takagi-Sugeno-Kang
(TSK) fuzzy system (MVRL_FS). The method realizes multi-view representation
learning from two aspects. First, multi-view data are transformed into a
high-dimensional fuzzy feature space, while the common information between
views and specific information of each view are explored simultaneously.
Second, a new regularization method based on L_(2,1)-norm regression is
proposed to mine the consistency information between views, while the geometric
structure of the data is preserved through the Laplacian graph. Finally,
extensive experiments on many benchmark multi-view datasets are conducted to
validate the superiority of the proposed method.Comment: This work has been accepted by IEEE Transactions on Knowledge and
Data Engineerin
Scalable Incomplete Multi-View Clustering with Structure Alignment
The success of existing multi-view clustering (MVC) relies on the assumption
that all views are complete. However, samples are usually partially available
due to data corruption or sensor malfunction, which raises the research of
incomplete multi-view clustering (IMVC). Although several anchor-based IMVC
methods have been proposed to process the large-scale incomplete data, they
still suffer from the following drawbacks: i) Most existing approaches neglect
the inter-view discrepancy and enforce cross-view representation to be
consistent, which would corrupt the representation capability of the model; ii)
Due to the samples disparity between different views, the learned anchor might
be misaligned, which we referred as the Anchor-Unaligned Problem for Incomplete
data (AUP-ID). Such the AUP-ID would cause inaccurate graph fusion and degrades
clustering performance. To tackle these issues, we propose a novel incomplete
anchor graph learning framework termed Scalable Incomplete Multi-View
Clustering with Structure Alignment (SIMVC-SA). Specially, we construct the
view-specific anchor graph to capture the complementary information from
different views. In order to solve the AUP-ID, we propose a novel structure
alignment module to refine the cross-view anchor correspondence. Meanwhile, the
anchor graph construction and alignment are jointly optimized in our unified
framework to enhance clustering quality. Through anchor graph construction
instead of full graphs, the time and space complexity of the proposed SIMVC-SA
is proven to be linearly correlated with the number of samples. Extensive
experiments on seven incomplete benchmark datasets demonstrate the
effectiveness and efficiency of our proposed method. Our code is publicly
available at https://github.com/wy1019/SIMVC-SA
- …