46 research outputs found
Weisfeiler and Leman go Hyperbolic: Learning Distance Preserving Node Representations
In recent years, graph neural networks (GNNs) have emerged as a promising
tool for solving machine learning problems on graphs. Most GNNs are members of
the family of message passing neural networks (MPNNs). There is a close
connection between these models and the Weisfeiler-Leman (WL) test of
isomorphism, an algorithm that can successfully test isomorphism for a broad
class of graphs. Recently, much research has focused on measuring the
expressive power of GNNs. For instance, it has been shown that standard MPNNs
are at most as powerful as WL in terms of distinguishing non-isomorphic graphs.
However, these studies have largely ignored the distances between the
representations of nodes/graphs which are of paramount importance for learning
tasks. In this paper, we define a distance function between nodes which is
based on the hierarchy produced by the WL algorithm, and propose a model that
learns representations which preserve those distances between nodes. Since the
emerging hierarchy corresponds to a tree, to learn these representations, we
capitalize on recent advances in the field of hyperbolic neural networks. We
empirically evaluate the proposed model on standard node and graph
classification datasets where it achieves competitive performance with
state-of-the-art models
What Do GNNs Actually Learn? Towards Understanding their Representations
In recent years, graph neural networks (GNNs) have achieved great success in
the field of graph representation learning. Although prior work has shed light
into the expressiveness of those models (\ie whether they can distinguish pairs
of non-isomorphic graphs), it is still not clear what structural information is
encoded into the node representations that are learned by those models. In this
paper, we investigate which properties of graphs are captured purely by these
models, when no node attributes are available. Specifically, we study four
popular GNN models, and we show that two of them embed all nodes into the same
feature vector, while the other two models generate representations that are
related to the number of walks over the input graph. Strikingly, structurally
dissimilar nodes can have similar representations at some layer , if they
have the same number of walks of length . We empirically verify our
theoretical findings on real datasets
k-hop Graph Neural Networks
Graph neural networks (GNNs) have emerged recently as a powerful architecture
for learning node and graph representations. Standard GNNs have the same
expressive power as the Weisfeiler-Leman test of graph isomorphism in terms of
distinguishing non-isomorphic graphs. However, it was recently shown that this
test cannot identify fundamental graph properties such as connectivity and
triangle freeness. We show that GNNs also suffer from the same limitation. To
address this limitation, we propose a more expressive architecture, k-hop GNNs,
which updates a node's representation by aggregating information not only from
its direct neighbors, but from its k-hop neighborhood. We show that the
proposed architecture can identify fundamental graph properties. We evaluate
the proposed architecture on standard node classification and graph
classification datasets. Our experimental evaluation confirms our theoretical
findings since the proposed model achieves performance better or comparable to
standard GNNs and to state-of-the-art algorithms.Comment: Accepted at Neural Network
Graph Classification with 2D Convolutional Neural Networks
Graph learning is currently dominated by graph kernels, which, while
powerful, suffer some significant limitations. Convolutional Neural Networks
(CNNs) offer a very appealing alternative, but processing graphs with CNNs is
not trivial. To address this challenge, many sophisticated extensions of CNNs
have recently been introduced. In this paper, we reverse the problem: rather
than proposing yet another graph CNN model, we introduce a novel way to
represent graphs as multi-channel image-like structures that allows them to be
handled by vanilla 2D CNNs. Experiments reveal that our method is more accurate
than state-of-the-art graph kernels and graph CNNs on 4 out of 6 real-world
datasets (with and without continuous node attributes), and close elsewhere.
Our approach is also preferable to graph kernels in terms of time complexity.
Code and data are publicly available.Comment: Published at ICANN 201
Supervised Attention Using Homophily in Graph Neural Networks
Graph neural networks have become the standard approach for dealing with
learning problems on graphs. Among the different variants of graph neural
networks, graph attention networks (GATs) have been applied with great success
to different tasks. In the GAT model, each node assigns an importance score to
its neighbors using an attention mechanism. However, similar to other graph
neural networks, GATs aggregate messages from nodes that belong to different
classes, and therefore produce node representations that are not well separated
with respect to the different classes, which might hurt their performance. In
this work, to alleviate this problem, we propose a new technique that can be
incorporated into any graph attention model to encourage higher attention
scores between nodes that share the same class label. We evaluate the proposed
method on several node classification datasets demonstrating increased
performance over standard baseline models.Comment: Accepted at ICANN 202
Kernel Graph Convolutional Neural Networks
Graph kernels have been successfully applied to many graph classification
problems. Typically, a kernel is first designed, and then an SVM classifier is
trained based on the features defined implicitly by this kernel. This two-stage
approach decouples data representation from learning, which is suboptimal. On
the other hand, Convolutional Neural Networks (CNNs) have the capability to
learn their own features directly from the raw data during training.
Unfortunately, they cannot handle irregular data such as graphs. We address
this challenge by using graph kernels to embed meaningful local neighborhoods
of the graphs in a continuous vector space. A set of filters is then convolved
with these patches, pooled, and the output is then passed to a feedforward
network. With limited parameter tuning, our approach outperforms strong
baselines on 7 out of 10 benchmark datasets.Comment: Accepted at ICANN '1
Message Passing Attention Networks for Document Understanding
Graph neural networks have recently emerged as a very effective framework for
processing graph-structured data. These models have achieved state-of-the-art
performance in many tasks. Most graph neural networks can be described in terms
of message passing, vertex update, and readout functions. In this paper, we
represent documents as word co-occurrence networks and propose an application
of the message passing framework to NLP, the Message Passing Attention network
for Document understanding (MPAD). We also propose several hierarchical
variants of MPAD. Experiments conducted on 10 standard text classification
datasets show that our architectures are competitive with the state-of-the-art.
Ablation studies reveal further insights about the impact of the different
components on performance. Code is publicly available at:
https://github.com/giannisnik/mpad .Comment: Accepted at AAAI'2