What leads to generalization of object proposals?Download PDF

Published: 29 Jul 2020, Last Modified: 05 May 2023VIPriors PosterReaders: Everyone
Keywords: object proposals, object detection, generalization, weakly supervised
Supplementary Material: zip
TL;DR: Scaling up object detection in a more data-efficient way? We provide a quantitative analysis on dataset properties and modeling choices for good proposals for unseen classes
Abstract: Object proposal generation is often the first step in many detection models. It is lucrative to train a good proposal model, that generalizes to unseen classes. This could help scaling detection models to larger number of classes with fewer annotations. Motivated by this, we study how a detection model trained on a small set of source classes can provide proposals that generalize to unseen classes. We systematically study the properties of the dataset – visual diversity and label space granularity – required for good generalization. We show the trade-off between using fine-grained labels and coarse labels. We introduce the idea of prototypical classes: a set of sufficient and necessary classes required to train a detection model to obtain generalized proposals in a more data-efficient way. On the Open Images V4 dataset, we show that only 25% of the classes can be selected to form such a prototypical set. The resulting proposals from a model trained with these classes is only 4.3% worse than using all the classes, in terms of average recall (AR). We also demonstrate that Faster R-CNN model leads to better generalization of proposals compared to a single-stage network like RetinaNet.
3 Replies

Loading