30 research outputs found
Graph Convolutional Matrix Completion
We consider matrix completion for recommender systems from the point of view
of link prediction on graphs. Interaction data such as movie ratings can be
represented by a bipartite user-item graph with labeled edges denoting observed
ratings. Building on recent progress in deep learning on graph-structured data,
we propose a graph auto-encoder framework based on differentiable message
passing on the bipartite interaction graph. Our model shows competitive
performance on standard collaborative filtering benchmarks. In settings where
complimentary feature information or structured data such as a social network
is available, our framework outperforms recent state-of-the-art methods.Comment: 9 pages, 3 figures, updated with additional experimental evaluatio
DeepInf: Social Influence Prediction with Deep Learning
Social and information networking activities such as on Facebook, Twitter,
WeChat, and Weibo have become an indispensable part of our everyday life, where
we can easily access friends' behaviors and are in turn influenced by them.
Consequently, an effective social influence prediction for each user is
critical for a variety of applications such as online recommendation and
advertising.
Conventional social influence prediction approaches typically design various
hand-crafted rules to extract user- and network-specific features. However,
their effectiveness heavily relies on the knowledge of domain experts. As a
result, it is usually difficult to generalize them into different domains.
Inspired by the recent success of deep neural networks in a wide range of
computing applications, we design an end-to-end framework, DeepInf, to learn
users' latent feature representation for predicting social influence. In
general, DeepInf takes a user's local network as the input to a graph neural
network for learning her latent social representation. We design strategies to
incorporate both network structures and user-specific features into
convolutional neural and attention networks. Extensive experiments on Open
Academic Graph, Twitter, Weibo, and Digg, representing different types of
social and information networks, demonstrate that the proposed end-to-end
model, DeepInf, significantly outperforms traditional feature engineering-based
approaches, suggesting the effectiveness of representation learning for social
applications.Comment: 10 pages, 5 figures, to appear in KDD 2018 proceeding
MGCN: Semi-supervised Classification in Multi-layer Graphs with Graph Convolutional Networks
Graph embedding is an important approach for graph analysis tasks such as
node classification and link prediction. The goal of graph embedding is to find
a low dimensional representation of graph nodes that preserves the graph
information. Recent methods like Graph Convolutional Network (GCN) try to
consider node attributes (if available) besides node relations and learn node
embeddings for unsupervised and semi-supervised tasks on graphs. On the other
hand, multi-layer graph analysis has been received attention recently. However,
the existing methods for multi-layer graph embedding cannot incorporate all
available information (like node attributes). Moreover, most of them consider
either type of nodes or type of edges, and they do not treat within and between
layer edges differently. In this paper, we propose a method called MGCN that
utilizes the GCN for multi-layer graphs. MGCN embeds nodes of multi-layer
graphs using both within and between layers relations and nodes attributes. We
evaluate our method on the semi-supervised node classification task.
Experimental results demonstrate the superiority of the proposed method to
other multi-layer and single-layer competitors and also show the positive
effect of using cross-layer edges
Competence of graph convolutional network in anti-money laundering in Bitcoin Blockchain
Graph networks are extensively used as an essential framework to analyse the interconnections between transactions and capture illicit behaviour in Bitcoin blockchain. Due to the complexity of Bitcoin transaction graph, the prediction of illicit transactions has become a challenging problem to unveil illicit services over the network. Graph Convolutional Network, a graph neural network based spectral approach, has recently emerged and gained much attention regarding graph-structured data. Previous research has highlighted the degraded performance of the latter approach to predict illicit transactions using, a Bitcoin transaction graph, so-called Elliptic data derived from Bitcoin blockchain. Motivated by the previous work, we seek to explore graph convolutions in a novel way. For this purpose, we present a novel approach that is modelled using the existing Graph Convolutional Network intertwined with linear layers. Concisely, we concatenate node embeddings obtained from graph convolutional layers with a single hidden layer derived from the linear transformation of the node feature matrix and followed by Multi-layer Perceptron. Our approach is evaluated using Elliptic data, wherein efficient accuracy is yielded. The proposed approach outperforms the original work of same data set
Contextual Graph Attention for Answering Logical Queries over Incomplete Knowledge Graphs
Recently, several studies have explored methods for using KG embedding to
answer logical queries. These approaches either treat embedding learning and
query answering as two separated learning tasks, or fail to deal with the
variability of contributions from different query paths. We proposed to
leverage a graph attention mechanism to handle the unequal contribution of
different query paths. However, commonly used graph attention assumes that the
center node embedding is provided, which is unavailable in this task since the
center node is to be predicted. To solve this problem we propose a multi-head
attention-based end-to-end logical query answering model, called Contextual
Graph Attention model(CGA), which uses an initial neighborhood aggregation
layer to generate the center embedding, and the whole model is trained jointly
on the original KG structure as well as the sampled query-answer pairs. We also
introduce two new datasets, DB18 and WikiGeo19, which are rather large in size
compared to the existing datasets and contain many more relation types, and use
them to evaluate the performance of the proposed model. Our result shows that
the proposed CGA with fewer learnable parameters consistently outperforms the
baseline models on both datasets as well as Bio dataset.Comment: 8 pages, 3 figures, camera ready version of article accepted to K-CAP
2019, Marina del Rey, California, United State
Poet: Product-oriented Video Captioner for E-commerce
In e-commerce, a growing number of user-generated videos are used for product
promotion. How to generate video descriptions that narrate the user-preferred
product characteristics depicted in the video is vital for successful
promoting. Traditional video captioning methods, which focus on routinely
describing what exists and happens in a video, are not amenable for
product-oriented video captioning. To address this problem, we propose a
product-oriented video captioner framework, abbreviated as Poet. Poet firstly
represents the videos as product-oriented spatial-temporal graphs. Then, based
on the aspects of the video-associated product, we perform knowledge-enhanced
spatial-temporal inference on those graphs for capturing the dynamic change of
fine-grained product-part characteristics. The knowledge leveraging module in
Poet differs from the traditional design by performing knowledge filtering and
dynamic memory modeling. We show that Poet achieves consistent performance
improvement over previous methods concerning generation quality, product
aspects capturing, and lexical diversity. Experiments are performed on two
product-oriented video captioning datasets, buyer-generated fashion video
dataset (BFVD) and fan-generated fashion video dataset (FFVD), collected from
Mobile Taobao. We will release the desensitized datasets to promote further
investigations on both video captioning and general video analysis problems.Comment: 10 pages, 3 figures, to appear in ACM MM 2020 proceeding
Relation Extraction with Self-determined Graph Convolutional Network
Relation Extraction is a way of obtaining the semantic relationship between
entities in text. The state-of-the-art methods use linguistic tools to build a
graph for the text in which the entities appear and then a Graph Convolutional
Network (GCN) is employed to encode the pre-built graphs. Although their
performance is promising, the reliance on linguistic tools results in a non
end-to-end process. In this work, we propose a novel model, the Self-determined
Graph Convolutional Network (SGCN), which determines a weighted graph using a
self-attention mechanism, rather using any linguistic tool. Then, the
self-determined graph is encoded using a GCN. We test our model on the TACRED
dataset and achieve the state-of-the-art result. Our experiments show that SGCN
outperforms the traditional GCN, which uses dependency parsing tools to build
the graph.Comment: CIKM-202
Uncertainty-based Traffic Accident Anticipation with Spatio-Temporal Relational Learning
Traffic accident anticipation aims to predict accidents from dashcam videos
as early as possible, which is critical to safety-guaranteed self-driving
systems. With cluttered traffic scenes and limited visual cues, it is of great
challenge to predict how long there will be an accident from early observed
frames. Most existing approaches are developed to learn features of
accident-relevant agents for accident anticipation, while ignoring the features
of their spatial and temporal relations. Besides, current deterministic deep
neural networks could be overconfident in false predictions, leading to high
risk of traffic accidents caused by self-driving systems. In this paper, we
propose an uncertainty-based accident anticipation model with spatio-temporal
relational learning. It sequentially predicts the probability of traffic
accident occurrence with dashcam videos. Specifically, we propose to take
advantage of graph convolution and recurrent networks for relational feature
learning, and leverage Bayesian neural networks to address the intrinsic
variability of latent relational representations. The derived uncertainty-based
ranking loss is found to significantly boost model performance by improving the
quality of relational features. In addition, we collect a new Car Crash Dataset
(CCD) for traffic accident anticipation which contains environmental attributes
and accident reasons annotations. Experimental results on both public and the
newly-compiled datasets show state-of-the-art performance of our model. Our
code and CCD dataset are available at https://github.com/Cogito2012/UString.Comment: Accepted by ACM MM 202
Understanding Negative Sampling in Graph Representation Learning
Graph representation learning has been extensively studied in recent years.
Despite its potential in generating continuous embeddings for various networks,
both the effectiveness and efficiency to infer high-quality representations
toward large corpus of nodes are still challenging. Sampling is a critical
point to achieve the performance goals. Prior arts usually focus on sampling
positive node pairs, while the strategy for negative sampling is left
insufficiently explored. To bridge the gap, we systematically analyze the role
of negative sampling from the perspectives of both objective and risk,
theoretically demonstrating that negative sampling is as important as positive
sampling in determining the optimization objective and the resulted variance.
To the best of our knowledge, we are the first to derive the theory and
quantify that the negative sampling distribution should be positively but
sub-linearly correlated to their positive sampling distribution. With the
guidance of the theory, we propose MCNS, approximating the positive
distribution with self-contrast approximation and accelerating negative
sampling by Metropolis-Hastings. We evaluate our method on 5 datasets that
cover extensive downstream graph learning tasks, including link prediction,
node classification and personalized recommendation, on a total of 19
experimental settings. These relatively comprehensive experimental results
demonstrate its robustness and superiorities.Comment: KDD 202
XGNN: Towards Model-Level Explanations of Graph Neural Networks
Graphs neural networks (GNNs) learn node features by aggregating and
combining neighbor information, which have achieved promising performance on
many graph tasks. However, GNNs are mostly treated as black-boxes and lack
human intelligible explanations. Thus, they cannot be fully trusted and used in
certain application domains if GNN models cannot be explained. In this work, we
propose a novel approach, known as XGNN, to interpret GNNs at the model-level.
Our approach can provide high-level insights and generic understanding of how
GNNs work. In particular, we propose to explain GNNs by training a graph
generator so that the generated graph patterns maximize a certain prediction of
the model.We formulate the graph generation as a reinforcement learning task,
where for each step, the graph generator predicts how to add an edge into the
current graph. The graph generator is trained via a policy gradient method
based on information from the trained GNNs. In addition, we incorporate several
graph rules to encourage the generated graphs to be valid. Experimental results
on both synthetic and real-world datasets show that our proposed methods help
understand and verify the trained GNNs. Furthermore, our experimental results
indicate that the generated graphs can provide guidance on how to improve the
trained GNNs