Abstract: Highlights•We used attribution analysis to find that some transformer based SR methods can only utilize limited spatial range information during the reconstruction process.•To address this, we introduce Spatial Shuffle Multi-Head Self-attention (SS-MSA) for efficient global pixel dependency modeling and a local perceptual unit to enhance local feature information.•Our method surpasses existing approaches in reconstruction accuracy and visual performance across five benchmarks. Moreover, it reduces parameters by 40%, GPU memory by 30%, and inference time by 30% compared to transformer-based methods.
Loading