New Features and Insights for Pedestrian Detection
Abstract
Despite impressive progress in people detection the performance on challenging datasets like Caltech Pedestrians or TUD-Brussels is still unsatisfactory. In this work we show that motion features derived from optic flow yield substantial improvements on image sequences, if implemented correctly - even in the case of low-quality video and consequently degraded flow fields. Furthermore, we introduce a new feature, self-similarity on color channels, which consistently improves detection performance both for static images and for video sequences, across different datasets. In combination with HOG, these two features outperform the state-of-the-art by up to 20%. Finally, we report two insights concerning detector evaluations, which apply to classifier-based object detection in general. First, we show that a commonly under-estimated detail of training, the number of bootstrapping rounds, has a drastic influence on the relative (and absolute) performance of different feature/classifier combinations. Second, we discuss important intricacies of detector evaluation and show that current benchmarking protocols lack crucial details, which can distort evaluations.
Cite
Text
Walk et al. "New Features and Insights for Pedestrian Detection." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2010. doi:10.1109/CVPR.2010.5540102Markdown
[Walk et al. "New Features and Insights for Pedestrian Detection." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2010.](https://mlanthology.org/cvpr/2010/walk2010cvpr-new/) doi:10.1109/CVPR.2010.5540102BibTeX
@inproceedings{walk2010cvpr-new,
title = {{New Features and Insights for Pedestrian Detection}},
author = {Walk, Stefan and Majer, Nikodem and Schindler, Konrad and Schiele, Bernt},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year = {2010},
pages = {1030-1037},
doi = {10.1109/CVPR.2010.5540102},
url = {https://mlanthology.org/cvpr/2010/walk2010cvpr-new/}
}