Accelerating Molecular Graph Neural Networks via Knowledge Distillation

Published: 18 Nov 2023, Last Modified: 25 Nov 2023LoG 2023 OralEveryoneRevisionsBibTeX
Keywords: GNN, graph neural networks, knowledge distillation, molecules, molecular simulations
TL;DR: We use knowledge distillation to improve the predictive performance of different molecular GNN with zero additional computational cost at inference.
Abstract: Recent advances in graph neural networks (GNNs) have enabled a more comprehensive representation of molecules and molecular systems, thereby enhancing the precision of molecular property prediction and molecular simulations. Nonetheless, as the field has been progressing to bigger and more complex architectures, state-of-the-art GNNs have become largely prohibitive for many large-scale applications. In this paper, we explore the utility of knowledge distillation (KD) for accelerating molecular GNNs. To this end, we devise KD strategies that facilitate the distillation of hidden representations in directional and equivariant GNNs, and evaluate their performance on the regression task of energy and force prediction. We validate our protocols across different teacher-student configurations and datasets, and demonstrate that they can consistently boost the predictive accuracy of student models without any modifications to their architecture. All in all, we manage to close the gap in predictive accuracy between teacher and student models by as much as 96.7 % and 62.5 % for energy and force prediction respectively, while fully preserving the inference throughput of the more lightweight models.
Submission Type: Extended abstract (max 4 main pages).
Poster: png
Poster Preview: png
Submission Number: 190
Loading