Graph Autoencoder for Graph Compression and Representation LearningDownload PDF

Mar 04, 2021 (edited Apr 20, 2021)Neural Compression Workshop @ ICLR 2021Readers: Everyone
  • Keywords: Graph Autoencoder, Graph Compression, Representation Learning, Graph Pooling
  • TL;DR: We propose a novel Graph Autoencoder structure, MIAGAE, which achieves state-of-the-art performance on graph compression and representation learning tasks.
  • Abstract: We consider the problem of graph data compression and representation. Recent developments in graph neural networks (GNNs) focus on generalizing convolutional neural networks (CNNs) to graph data, which includes redesigning convolution and pooling operations for graphs. However, few methods focus on effective graph compression to obtain a smaller graph, which can reconstruct the original full graph with less storage and can provide useful latent representations to improve downstream task performance. To fill this gap, we propose Multi-kernel Inductive Attention Graph Autoencoder (MIAGAE), which, instead of compressing nodes/edges separately, utilizes the node similarity and graph structure to compress all nodes and edges as a whole. Similarity attention graph pooling selects the most representative nodes with the most information by using the similarity and topology among nodes. Our multi-kernel Inductive-Convolution layer can focus on different aspects and learn more general node representations in evolving graphs. We demonstrate that MIAGAE outperforms state-of-the-art methods for graph compression and few-shot graph classification, with superior graph representation learning.
1 Reply