On-the-Fly Object Modeling While Tracking
Abstract
To implement a persistent tracker, we build a set of view-dependent object appearance models adoptively and automatically while tracking an object under different viewing angles. This collection of acquired models is indexed with respect to the view sphere. The acquired models aid recovery from tracking failure due to occlusion and changing view angle. In this paper, view-dependent object appearance is represented by intensity patches around detected Harris corners. The intensity patches from a model are matched to the current frame by solving a bipartite linear assignment problem with outlier exclusion and missed inlier recovery. Based on these reliable matches, the change in object rotation, translation and scale is estimated between consecutive frames using Procrustes analysis. The experimental results show good performance using a collection of view-specific patch-based models for detection and tracking of vehicles in low-resolution airborne video.
Cite
Text
Yin and Collins. "On-the-Fly Object Modeling While Tracking." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007. doi:10.1109/CVPR.2007.383237Markdown
[Yin and Collins. "On-the-Fly Object Modeling While Tracking." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007.](https://mlanthology.org/cvpr/2007/yin2007cvpr-fly/) doi:10.1109/CVPR.2007.383237BibTeX
@inproceedings{yin2007cvpr-fly,
title = {{On-the-Fly Object Modeling While Tracking}},
author = {Yin, Zhaozheng and Collins, Robert T.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year = {2007},
doi = {10.1109/CVPR.2007.383237},
url = {https://mlanthology.org/cvpr/2007/yin2007cvpr-fly/}
}