Journal
NEURAL NETWORKS
Volume 142, Issue -, Pages 1-19Publisher
PERGAMON-ELSEVIER SCIENCE LTD
DOI: 10.1016/j.neunet.2021.04.015
Keywords
Graph autoencoders; Graph variational autoencoders; Scalability; Graph convolutional networks; Link prediction; Node clustering
Ask authors/readers for more resources
This paper introduces FastGAE, a general framework to scale graph AE and VAE to large graphs, which significantly speeds up training and improves performance through an effective stochastic subgraph decoding scheme. FastGAE demonstrates effectiveness on various real-world graphs, outperforming existing approaches by a wide margin.
Graph autoencoders (AE) and variational autoencoders (VAE) are powerful node embedding methods, but suffer from scalability issues. In this paper, we introduce FastGAE, a general framework to scale graph AE and VAE to large graphs with millions of nodes and edges. Our strategy, based on an effective stochastic subgraph decoding scheme, significantly speeds up the training of graph AE and VAE while preserving or even improving performances. We demonstrate the effectiveness of FastGAE on various real-world graphs, outperforming the few existing approaches to scale graph AE and VAE by a wide margin. (C) 2021 Elsevier Ltd. All rights reserved.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available