DyGNN: Algorithm and Architecture Support of Dynamic Pruning for Graph Neural Networks

Published: 01 Jan 2021, Last Modified: 11 Apr 2025DAC 2021EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Recently, graph neural networks (GNNs) have achieved great success for graph representation learning tasks. Enlightened by the fact that numerous message passing redundancies exist in GNNs, we propose DyGNN, which speeds up GNNs by reducing redundancies. DyGNN is supported by an algorithm and architecture co-design. The proposed algorithm can dynamically prune vertices and edges during execution without accuracy loss. An architecture is designed to support dynamic pruning and transform it into performance improvement. DyGNN opens new directions for accelerating GNNs by pruning vertices and edges. DyGNN gains average $2\times$ speedup with accuracy improvement of 4% compared with state-of-the-art GNN accelerators.
Loading