Referring Expression Counting
Abstract
Existing counting tasks are limited to the class level which don't account for fine-grained details within the class. In real applications it often requires in-context or referring human input for counting target objects. Take urban analysis as an example fine-grained information such as traffic flow in different directions pedestrians and vehicles waiting or moving at different sides of the junction is more beneficial. Current settings of both class-specific and class-agnostic counting treat objects of the same class indifferently which pose limitations in real use cases. To this end we propose a new task named Referring Expression Counting (REC) which aims to count objects with different attributes within the same class. To evaluate the REC task we create a novel dataset named REC-8K which contains 8011 images and 17122 referring expressions. Experiments on REC-8K show that our proposed method achieves state-of-the-art performance compared with several text-based counting methods and an open-set object detection model. We also outperform prior models on the class agnostic counting (CAC) benchmark [36] for the zero-shot setting and perform on par with the few-shot methods. Code and dataset is available at https://github.com/sydai/referring-expression-counting.
Cite
Text
Dai et al. "Referring Expression Counting." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.01607Markdown
[Dai et al. "Referring Expression Counting." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/dai2024cvpr-referring/) doi:10.1109/CVPR52733.2024.01607BibTeX
@inproceedings{dai2024cvpr-referring,
title = {{Referring Expression Counting}},
author = {Dai, Siyang and Liu, Jun and Cheung, Ngai-Man},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2024},
pages = {16985-16995},
doi = {10.1109/CVPR52733.2024.01607},
url = {https://mlanthology.org/cvpr/2024/dai2024cvpr-referring/}
}