Harmformer: Harmonic Networks Meet Transformers for Continuous Roto-Translation Equivariance

Published: 23 Oct 2024, Last Modified: 24 Feb 2025NeurReps 2024 OralEveryoneRevisionsBibTeXCC BY 4.0
Keywords: transformers, roto-translation, robustness, invariance, equivariance, geometric deep learning
TL;DR: Transformer Equivariant to Continuous Roto-Translation in 2D
Abstract: Convolutional Neural Networks exhibit inherent equivariance to image translation, leading to efficient parameter and data usage, faster learning, and improved robustness. The concept of translation equivariant networks has been successfully extended to rotation transformation using group convolution for discrete rotation groups and harmonic functions for the continuous rotation group encompassing $360^\circ$. We explore the compatibility of the Self-Attention mechanism with full rotation equivariance, in contrast to previous studies that focused on discrete rotation. We introduce the Harmformer, a harmonic transformer with a convolutional stem that achieves equivariance for both translation and continuous rotation. Accompanied by an end-to-end equivariance proof, the Harmformer not only outperforms previous equivariant transformers, but also demonstrates inherent stability under any continuous rotation, even without seeing rotated samples during training.
Submission Number: 5
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview