Person Search via a Mask-Guided Two-Stream CNN Model
Abstract
In this work, we tackle the problem of person search, which is a challenging task consisted of pedestrian detection and person re-identification~(re-ID). Instead of sharing representations in a single joint model, we find that separating detector and re-ID feature extraction yields better performance. In order to extract more representative features for each identity, we segment out the foreground person from the original image patch. We propose a simple yet effective re-ID method, which models foreground person and original image patches individually, and obtains enriched representations from two separate CNN streams. From the experiments on two standard person search benchmarks of CUHK-SYSU and PRW, we achieve mAP of $83.0%$ and $32.6%$ respectively, surpassing the state of the art by a large margin (more than 5pp).
Cite
Text
Chen et al. "Person Search via a Mask-Guided Two-Stream CNN Model." Proceedings of the European Conference on Computer Vision (ECCV), 2018. doi:10.1007/978-3-030-01234-2_45Markdown
[Chen et al. "Person Search via a Mask-Guided Two-Stream CNN Model." Proceedings of the European Conference on Computer Vision (ECCV), 2018.](https://mlanthology.org/eccv/2018/chen2018eccv-person/) doi:10.1007/978-3-030-01234-2_45BibTeX
@inproceedings{chen2018eccv-person,
title = {{Person Search via a Mask-Guided Two-Stream CNN Model}},
author = {Chen, Di and Zhang, Shanshan and Ouyang, Wanli and Yang, Jian and Tai, Ying},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2018},
doi = {10.1007/978-3-030-01234-2_45},
url = {https://mlanthology.org/eccv/2018/chen2018eccv-person/}
}