Need for Speed: A Benchmark for Higher Frame Rate Object Tracking

Abstract

In this paper, we propose the first higher frame rate video dataset (called Need for Speed - NfS) and benchmark for visual object tracking. The dataset consists of 100 videos (380K frames) captured with now commonly available higher frame rate (240 FPS) cameras from real world scenarios. All frames are annotated with axis aligned bounding boxes and all sequences are manually labelled with nine visual attributes - such as occlusion, fast motion, background clutter, etc. Our benchmark provides an extensive evaluation of many recent and state-of-the-art trackers on higher frame rate sequences. We ranked each of these trackers according to their tracking accuracy and real-time performance. One of our surprising conclusions is that at higher frame rates, simple trackers such as correlation filters outperform complex methods based on deep networks. This suggests that for practical applications (such as in robotics or embedded vision), one needs to carefully tradeoff bandwidth constraints associated with higher frame rate acquisition, computational costs of real-time analysis, and the required application accuracy. Our dataset and benchmark allows for the first time (to our knowledge) systematic exploration of such issues, and will be made available to allow for further research in this space.

Cite

Text

Galoogahi et al. "Need for Speed: A Benchmark for Higher Frame Rate Object Tracking." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.128

Markdown

[Galoogahi et al. "Need for Speed: A Benchmark for Higher Frame Rate Object Tracking." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/galoogahi2017iccv-need/) doi:10.1109/ICCV.2017.128

BibTeX

@inproceedings{galoogahi2017iccv-need,
  title     = {{Need for Speed: A Benchmark for Higher Frame Rate Object Tracking}},
  author    = {Galoogahi, Hamed Kiani and Fagg, Ashton and Huang, Chen and Ramanan, Deva and Lucey, Simon},
  booktitle = {International Conference on Computer Vision},
  year      = {2017},
  doi       = {10.1109/ICCV.2017.128},
  url       = {https://mlanthology.org/iccv/2017/galoogahi2017iccv-need/}
}