112,132 research outputs found
Multi-view Graph Embedding with Hub Detection for Brain Network Analysis
Multi-view graph embedding has become a widely studied problem in the area of
graph learning. Most of the existing works on multi-view graph embedding aim to
find a shared common node embedding across all the views of the graph by
combining the different views in a specific way. Hub detection, as another
essential topic in graph mining has also drawn extensive attentions in recent
years, especially in the context of brain network analysis. Both the graph
embedding and hub detection relate to the node clustering structure of graphs.
The multi-view graph embedding usually implies the node clustering structure of
the graph based on the multiple views, while the hubs are the boundary-spanning
nodes across different node clusters in the graph and thus may potentially
influence the clustering structure of the graph. However, none of the existing
works in multi-view graph embedding considered the hubs when learning the
multi-view embeddings. In this paper, we propose to incorporate the hub
detection task into the multi-view graph embedding framework so that the two
tasks could benefit each other. Specifically, we propose an auto-weighted
framework of Multi-view Graph Embedding with Hub Detection (MVGE-HD) for brain
network analysis. The MVGE-HD framework learns a unified graph embedding across
all the views while reducing the potential influence of the hubs on blurring
the boundaries between node clusters in the graph, thus leading to a clear and
discriminative node clustering structure for the graph. We apply MVGE-HD on two
real multi-view brain network datasets (i.e., HIV and Bipolar). The
experimental results demonstrate the superior performance of the proposed
framework in brain network analysis for clinical investigation and application
MEGAN: A Generative Adversarial Network for Multi-View Network Embedding
Data from many real-world applications can be naturally represented by
multi-view networks where the different views encode different types of
relationships (e.g., friendship, shared interests in music, etc.) between
real-world individuals or entities. There is an urgent need for methods to
obtain low-dimensional, information preserving and typically nonlinear
embeddings of such multi-view networks. However, most of the work on multi-view
learning focuses on data that lack a network structure, and most of the work on
network embeddings has focused primarily on single-view networks. Against this
background, we consider the multi-view network representation learning problem,
i.e., the problem of constructing low-dimensional information preserving
embeddings of multi-view networks. Specifically, we investigate a novel
Generative Adversarial Network (GAN) framework for Multi-View Network
Embedding, namely MEGAN, aimed at preserving the information from the
individual network views, while accounting for connectivity across (and hence
complementarity of and correlations between) different views. The results of
our experiments on two real-world multi-view data sets show that the embeddings
obtained using MEGAN outperform the state-of-the-art methods on node
classification, link prediction and visualization tasks.Comment: Proceedings of the Twenty-Eighth International Joint Conference on
Artificial Intelligence, IJCAI-1
End-to-End Cross-Modality Retrieval with CCA Projections and Pairwise Ranking Loss
Cross-modality retrieval encompasses retrieval tasks where the fetched items
are of a different type than the search query, e.g., retrieving pictures
relevant to a given text query. The state-of-the-art approach to cross-modality
retrieval relies on learning a joint embedding space of the two modalities,
where items from either modality are retrieved using nearest-neighbor search.
In this work, we introduce a neural network layer based on Canonical
Correlation Analysis (CCA) that learns better embedding spaces by analytically
computing projections that maximize correlation. In contrast to previous
approaches, the CCA Layer (CCAL) allows us to combine existing objectives for
embedding space learning, such as pairwise ranking losses, with the optimal
projections of CCA. We show the effectiveness of our approach for
cross-modality retrieval on three different scenarios (text-to-image,
audio-sheet-music and zero-shot retrieval), surpassing both Deep CCA and a
multi-view network using freely learned projections optimized by a pairwise
ranking loss, especially when little training data is available (the code for
all three methods is released at: https://github.com/CPJKU/cca_layer).Comment: Preliminary version of a paper published in the International Journal
of Multimedia Information Retrieva
Multi-view Graph Convolutional Networks with Differentiable Node Selection
Multi-view data containing complementary and consensus information can
facilitate representation learning by exploiting the intact integration of
multi-view features. Because most objects in real world often have underlying
connections, organizing multi-view data as heterogeneous graphs is beneficial
to extracting latent information among different objects. Due to the powerful
capability to gather information of neighborhood nodes, in this paper, we apply
Graph Convolutional Network (GCN) to cope with heterogeneous-graph data
originating from multi-view data, which is still under-explored in the field of
GCN. In order to improve the quality of network topology and alleviate the
interference of noises yielded by graph fusion, some methods undertake sorting
operations before the graph convolution procedure. These GCN-based methods
generally sort and select the most confident neighborhood nodes for each
vertex, such as picking the top-k nodes according to pre-defined confidence
values. Nonetheless, this is problematic due to the non-differentiable sorting
operators and inflexible graph embedding learning, which may result in blocked
gradient computations and undesired performance. To cope with these issues, we
propose a joint framework dubbed Multi-view Graph Convolutional Network with
Differentiable Node Selection (MGCN-DNS), which is constituted of an adaptive
graph fusion layer, a graph learning module and a differentiable node selection
schema. MGCN-DNS accepts multi-channel graph-structural data as inputs and aims
to learn more robust graph fusion through a differentiable neural network. The
effectiveness of the proposed method is verified by rigorous comparisons with
considerable state-of-the-art approaches in terms of multi-view semi-supervised
classification tasks
Graph Representation Learning-Based Recommender Systems
University of Technology Sydney. Faculty of Engineering and Information Technology.Personalized recommendation has been applied to many online services such as E-commerce and adverting. It facilitates users to discover a small set of relevant items, which meet their personalized interests, from many choices. Nowadays, various auxiliary information on users and items become increasingly available in online platforms, such as user demographics, social relations, and item knowledge. More recent evidences suggests that incorporating such auxiliary data with collaborative filtering can better capture the underlying and complex user-item relationships, and further achieve higher recommendation quality.
In this thesis, we focus on auxiliary data with graph structure, such as social networks and knowledge graphs (KG). For example, we can improve recommendation performance by mining social relationships between users, and also by using knowledge graphs to enhance the semantics of recommended items. Network representation learning aims to represent each vertex in a network (graph) as a low-dimensional vector while still preserving its structural information. Due to the availability of massive graph data in recommender systems, it is a promising approach to combine network representation learning with recommendation. Applying the learned graph features to recommender systems will effectively enhance the learning ability of the recommender systems and improve the accuracy and user satisfaction of the recommender systems. For network representation learning and its application in recommendation systems, the major contributions of this thesis are as follows:
(1) Attention-based Adversarial Autoencoder for Multi-scale Network Embedding. Existing Network representation methods usually adopt a one-size-fits-all approach when concerning multi-scale structure information, such as first- and second-order proximity of nodes, ignoring the fact that different scales play different roles in embedding learning. We propose an Attention-based Adversarial Autoencoder Network Embedding (AAANE) framework, which promotes the collaboration of different scales and lets them vote for robust representations.
(2) Multi-modal Multi-view Bayesian Semantic Embedding for Community Question Answering: Semantic embedding has demonstrated its value in latent representation learning of data, and can be effectively adopted for many applications. However, it is difficult to propose a joint learning framework for semantic embedding in Community Question Answer (CQA), because CQA data have multi-view and sparse properties. In this thesis, we propose a generic Multi-modal Multi-view Semantic Embedding (MMSE) framework via a Bayesian model for question answering.
(3) Context-Dependent Propagating-based Video Recommendation in Multi-modal Heterogeneous Information Networks. Conventional approaches to video recommendation primarily focus on exploiting content features or simple user-video interactions to model the users’ preferences. However these methods fail to model the complex video context interdependency, which is obscure/hidden in heterogeneous auxiliary data. In this paper, we propose a Context-Dependent Propagating Recommendation network (CDPRec) to obtain accurate video embedding and capture global context cues among videos in HINs. The CDPRec can iteratively propagate the contexts of a video along links in a graph-structured HIN and explore multiple types of dependencies among the surrounding video nodes.
(4) Knowledge Graph Enhanced Neural Collaborative Filtering. Existing neural collaborative filtering (NCF) recommendation methods suffer from severe sparsity problem. Knowledge Graph (KG), which commonly consists of fruitful connected facts about items, presents an unprecedented opportunity to alleviate the sparsity problem. However, NCF only methods can hardly model the high-order connectivity in KG, and ignores complex pairwise correlations between user/item embedding dimensions. To address these issues, we propose a novel Knowledge graph enhanced Neural Collaborative Recommendation (K-NCR) framework, which effectively combines user-item interaction information and auxiliary knowledge information for recommendation
- …