SparseSwin: Swin transformer with sparse transformer block

Published: 01 Jan 2024, Last Modified: 01 Oct 2024Neurocomputing 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•Proposing Sparse Transformer Block with limited latent tokens to enhance Swin Transformer efficiency and performance.•We examined the effect of regularization on the attention weight obtained in SparTa Block.•We obtained accuracy improvement on ImageNet100, CIFAR10, and CIFAR100 benchmark datasets.•We obtained a smaller parameter size compared to the state-of-the-art Transformer model.
Loading