Few-shot Class-agnostic Counting with Occlusion Augmentation and Localization

Published: 01 Jan 2024, Last Modified: 23 Aug 2024ISCAS 2024EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: Most existing few-shot class-agnostic counting (FCAC) methods follow the extract-and-compare pipeline to count all instances of an arbitrary category in the query image given a few exemplars. However, these methods generate the density map rather than the exact instance location for counting, which is less intuitive and accurate than the latter. Besides, how to alleviate the problem of occlusion is ignored in most existing work. To solve the above problems, this paper proposes an Occlusion-Augmented Localization Network (OALNet), which extracts multiple occluded features of exemplars for comparison and utilizes the precise position of instances for more accurate and confident counting results. Specifically, the OALNet is in an extract-and-attention manner. It includes an Occluded Feature Generation module to deal with the occlusion problem in query images. Besides, the OALNet adopts the Feature Attention module to improve the extracted feature by self-attention and model the relationship between the exemplar features and query features by cross-attention. Compared with other FCAC methods, experimental results demonstrate that the proposed OALNet achieves superior performance.
Loading