Graph neural networks (GNNs) work well when the graph structure is provided. However, this structure may not always be available in real-world applications. One solution to this problem is to infer a task-specific latent structure and then apply a GNN to the inferred graph. Unfortunately, the space of possible graph structures grows super-exponentially with the number of nodes and so the task-specific supervision may be insufficient for learning both the structure and the GNN parameters. In this work, we propose the Simultaneous Learning of Adjacency and GNN Parameters with Self-supervision, or SLAPS, a method that provides more supervision for inferring a graph structure through self-supervision. A comprehensive experimental study demonstrates that SLAPS scales to large graphs with hundreds of thousands of nodes and outperforms several models that have been proposed to learn a task-specific graph structure on established benchmarks.

Authors
BibTeX

@inproceedings{slaps,
title = {SLAPS: Self-Supervision Improves Structure Learning for Graph Neural Networks},
authors = {Bahare Fatemi and Layla El Asri and Seyed Mehran Kazemi},
year = {2021},
booktitle = {Proceedings of Advances in Neural Information Processing Systems (NeurIPS)},
}