Constituent Attention for Vision Transformers

Published: 01 Jan 2023, Last Modified: 28 Aug 2024Comput. Vis. Image Underst. 2023EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Highlights•Spatial constraint contributes to concentrated attention and performance gain in MSA.•Concentrated attention helps ViTs facilitate optimization at data-scarce cases.•Hierarchical constraint yields progressive attention across different layers.•Layer-wise reasoning facilitates the understanding of inner workings in ViTs.
Loading