2 research outputs found
Benchmarking Graph Neural Networks
Graph neural networks (GNNs) have become the standard toolkit for analyzing
and learning from data on graphs. As the field grows, it becomes critical to
identify key architectures and validate new ideas that generalize to larger,
more complex datasets. Unfortunately, it has been increasingly difficult to
gauge the effectiveness of new models in the absence of a standardized
benchmark with consistent experimental settings. In this paper, we introduce a
reproducible GNN benchmarking framework, with the facility for researchers to
add new models conveniently for arbitrary datasets. We demonstrate the
usefulness of our framework by presenting a principled investigation into the
recent Weisfeiler-Lehman GNNs (WL-GNNs) compared to message passing-based graph
convolutional networks (GCNs) for a variety of graph tasks, i.e. graph
regression/classification and node/link prediction, with medium-scale datasets.Comment: Benchmarking framework on GitHub at
https://github.com/graphdeeplearning/benchmarking-gnn