Scale-Invariant Implicit Neural Representations for Object Counting

Abstract

Existing object counting methods relying on Density Map Estimation(DME) struggle with large variations in object size or input image resolution due to different imaging conditions and perspective effects. Especially, discrete grid representations of density maps result in information loss with blurred or vanished details for low-resolution inputs. To overcome these limitations, we design new Scale-Invariant Implicit Neural Representations (SI-INR) for counting to map arbitrary-scale input signals into a continuous function space, where function values over continuous spatial coordinates indicate probabilities observing objects of interest. Extensive experiments on diverse benchmark datasets have validated that SI-INR achieves robust counting performances with respect to changing input sizes, leading to better or comparable object counting accuracy compared to state-of-the-art methods.

Cite

Text

Xu et al. "Scale-Invariant Implicit Neural Representations for Object Counting." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.

Markdown

[Xu et al. "Scale-Invariant Implicit Neural Representations for Object Counting." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2025.](https://mlanthology.org/cvprw/2025/xu2025cvprw-scaleinvariant/)

BibTeX

@inproceedings{xu2025cvprw-scaleinvariant,
  title     = {{Scale-Invariant Implicit Neural Representations for Object Counting}},
  author    = {Xu, Siyuan and Wang, Yucheng and Luo, Xihaier and Yoon, Byung-Jun and Qian, Xiaoning},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2025},
  pages     = {2308-2318},
  url       = {https://mlanthology.org/cvprw/2025/xu2025cvprw-scaleinvariant/}
}