Primphormer: Efficient Graph Transformers with Primal Representations

Published: 01 May 2025, Last Modified: 18 Jun 2025ICML 2025 posterEveryoneRevisionsBibTeXCC BY 4.0
TL;DR: We propose a primal-dual framework for efficient graph Transformers which reduces the quadratic complexity.
Abstract: Graph Transformers (GTs) have emerged as a promising approach for graph representation learning. Despite their successes, the quadratic complexity of GTs limits scalability on large graphs due to their pair-wise computations. To fundamentally reduce the computational burden of GTs, we propose a primal-dual framework that interprets the self-attention mechanism on graphs as a dual representation. Based on this framework, we develop Primphormer, an efficient GT that leverages a primal representation with linear complexity. Theoretical analysis reveals that Primphormer serves as a universal approximator for functions on both sequences and graphs, while also retaining its expressive power for distinguishing non-isomorphic graphs. Extensive experiments on various graph benchmarks demonstrate that Primphormer achieves competitive empirical results while maintaining a more user-friendly memory and computational costs.
Lay Summary: In this paper, we propose a method to reduce the heavy computation in graph Transformers. We investigate a new optimization problem whose dual representation coincides with the standard attention mechanism. This new optimization indicates a new representation in the primal space, where the heavy quadratic computation could be avoided, enhancing efficiency.
Primary Area: Deep Learning->Attention Mechanisms
Keywords: Efficient graph Transformer, self-attention mechanisms, primal representation, asymmetry, kernel methods.
Submission Number: 5859
Loading