How Expressive are Transformers in Spectral Domain for Graphs?

Published: 09 Jul 2022, Last Modified: 28 Feb 2023Accepted by TMLREveryoneRevisionsBibTeX
Abstract: The recent works proposing transformer-based models for graphs have proven the inadequacy of Vanilla Transformer for graph representation learning. To understand this inadequacy, there is a need to investigate if spectral analysis of the transformer will reveal insights into its expressive power. Similar studies already established that spectral analysis of Graph neural networks (GNNs) provides extra perspectives on their expressiveness. In this work, we systematically study and establish the link between the spatial and spectral domain in the realm of the transformer. We further provide a theoretical analysis that the spatial attention mechanism in the transformer cannot effectively capture the desired frequency response, thus, inherently limiting its expressiveness in spectral space. Therefore, we propose FeTA, a framework that aims to perform attention over the entire graph spectrum (i.e. actual frequency components of the graph) analogous to the attention in spatial space. Empirical results suggest that FeTA provides homogeneous performance gain against vanilla transformer across all tasks on standard benchmarks and can easily be extended to GNN-based models with low-pass characteristics (e.g., GAT).
Submission Length: Regular submission (no more than 12 pages of main content)
Changes Since Last Submission: Uploaded camera ready version.
Video: https://youtu.be/7JNDYQuRSas
Code: https://github.com/ansonb/FeTA_TMLR
Assigned Action Editor: ~Michal_Valko1
License: Creative Commons Attribution 4.0 International (CC BY 4.0)
Submission Number: 111
Loading