Abstract

Current top performing object detectors employ detection proposals to guide the search for objects, thereby avoiding exhaustive sliding window search across images. Despite the popularity and widespread use of detection proposals, it is unclear which trade-offs are made when using them during object detection. We provide an in-depth analysis of twelve proposal methods along with four baselines regarding proposal repeatability, ground truth annotation recall on PASCAL, ImageNet, and MS COCO, and their impact on DPM, R-CNN, and Fast R-CNN detection performance. Our analysis shows that for object detection improving proposal localisation accuracy is as important as improving recall. We introduce a novel metric, the average recall (AR), which rewards both high recall and good localisation and correlates surprisingly well with detection performance. Our findings show common strengths and weaknesses of existing methods, and provide insights and metrics for selecting and tuning proposal methods.

Keywords

Computer scienceObject detectionPascal (unit)RecallArtificial intelligencePrecision and recallMetric (unit)Ground truthPopularitySliding window protocolStrengths and weaknessesDetectorMachine learningPattern recognition (psychology)Data miningWindow (computing)

Affiliated Institutions

Related Publications

Publication Info

Year
2015
Type
article
Volume
38
Issue
4
Pages
814-830
Citations
749
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

749
OpenAlex
76
Influential
503
CrossRef

Cite This

Jan Hosang, Rodrigo Benenson, Piotr Dollár et al. (2015). What Makes for Effective Detection Proposals?. IEEE Transactions on Pattern Analysis and Machine Intelligence , 38 (4) , 814-830. https://doi.org/10.1109/tpami.2015.2465908

Identifiers

DOI
10.1109/tpami.2015.2465908
PMID
26959679
arXiv
1502.05082

Data Quality

Data completeness: 84%