In this paper, we present a general framework to scale graph autoencoders
(AE) and graph variational autoencoders (VAE). This framework leverages graph
degeneracy concepts to train models only from a dense subset of nodes instead
of using the entire graph. Together with a simple yet effective propagation
mechanism, our approach significantly improves scalability and training speed
while preserving performance. We evaluate and discuss our method on several
variants of existing graph AE and VAE, providing the first application of these
models to large graphs with up to millions of nodes and edges. We achieve
empirically competitive results w.r.t. several popular scalable node embedding
methods, which emphasizes the relevance of pursuing further research towards
more scalable graph AE and VAE.Comment: International Joint Conference on Artificial Intelligence (IJCAI
2019