On-the-Fly Object Modeling While Tracking

Abstract

To implement a persistent tracker, we build a set of view-dependent object appearance models adoptively and automatically while tracking an object under different viewing angles. This collection of acquired models is indexed with respect to the view sphere. The acquired models aid recovery from tracking failure due to occlusion and changing view angle. In this paper, view-dependent object appearance is represented by intensity patches around detected Harris corners. The intensity patches from a model are matched to the current frame by solving a bipartite linear assignment problem with outlier exclusion and missed inlier recovery. Based on these reliable matches, the change in object rotation, translation and scale is estimated between consecutive frames using Procrustes analysis. The experimental results show good performance using a collection of view-specific patch-based models for detection and tracking of vehicles in low-resolution airborne video.

Cite

Text

Yin and Collins. "On-the-Fly Object Modeling While Tracking." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007. doi:10.1109/CVPR.2007.383237

Markdown

[Yin and Collins. "On-the-Fly Object Modeling While Tracking." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007.](https://mlanthology.org/cvpr/2007/yin2007cvpr-fly/) doi:10.1109/CVPR.2007.383237

BibTeX

@inproceedings{yin2007cvpr-fly,
  title     = {{On-the-Fly Object Modeling While Tracking}},
  author    = {Yin, Zhaozheng and Collins, Robert T.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2007},
  doi       = {10.1109/CVPR.2007.383237},
  url       = {https://mlanthology.org/cvpr/2007/yin2007cvpr-fly/}
}