Learning to Count Everything
Abstract
Existing works on visual counting primarily focus on one specific category at a time, such as people, animals, and cells. In this paper, we are interested in counting everything, that is to count objects from any category given only a few annotated instances from that category. To this end, we pose counting as a few-shot regression task. To tackle this task, we present a novel method that takes a query image together with a few exemplar objects from the query image and predicts a density map for the presence of all objects of interest in the query image. We also present a novel adaptation strategy to adapt our network to any novel visual category at test time, using only a few exemplar objects from the novel category. We also introduce a dataset of 147 object categories containing over 6000 images that are suitable for the few-shot counting task. The images are annotated with two types of annotation, dots and bounding boxes, and they can be used for developing few-shot counting models. Experiments on this dataset shows that our method outperforms several state-of-the-art object detectors and few-shot counting approaches. Our code and dataset can be found at https://github.com/cvlab-stonybrook/LearningToCountEverything.
Cite
Text
Ranjan et al. "Learning to Count Everything." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00340Markdown
[Ranjan et al. "Learning to Count Everything." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/ranjan2021cvpr-learning/) doi:10.1109/CVPR46437.2021.00340BibTeX
@inproceedings{ranjan2021cvpr-learning,
title = {{Learning to Count Everything}},
author = {Ranjan, Viresh and Sharma, Udbhav and Nguyen, Thu and Hoai, Minh},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2021},
pages = {3394-3403},
doi = {10.1109/CVPR46437.2021.00340},
url = {https://mlanthology.org/cvpr/2021/ranjan2021cvpr-learning/}
}