MARS: A Video Benchmark for Large-Scale Person Re-Identification
Abstract
This paper considers person re-identification (re-id) in videos. We introduce a new video re-id dataset, named M otion A nalysis and R e-identification S et (MARS), a video extension of the Market-1501 dataset. To our knowledge, MARS is the largest video re-id dataset to date. Containing 1,261 IDs and around 20,000 tracklets, it provides rich visual information compared to image-based datasets. Meanwhile, MARS reaches a step closer to practice. The tracklets are automatically generated by the Deformable Part Model (DPM) as pedestrian detector and the GMMCP tracker. A number of false detection/tracking results are also included as distractors which would exist predominantly in practical video databases. Extensive evaluation of the state-of-the-art methods including the space-time descriptors and CNN is presented. We show that CNN in classification mode can be trained from scratch using the consecutive bounding boxes of each identity. The learned CNN embedding outperforms other competing methods considerably and has good generalization ability on other video re-id datasets upon fine-tuning.
Cite
Text
Zheng et al. "MARS: A Video Benchmark for Large-Scale Person Re-Identification." European Conference on Computer Vision, 2016. doi:10.1007/978-3-319-46466-4_52Markdown
[Zheng et al. "MARS: A Video Benchmark for Large-Scale Person Re-Identification." European Conference on Computer Vision, 2016.](https://mlanthology.org/eccv/2016/zheng2016eccv-mars/) doi:10.1007/978-3-319-46466-4_52BibTeX
@inproceedings{zheng2016eccv-mars,
title = {{MARS: A Video Benchmark for Large-Scale Person Re-Identification}},
author = {Zheng, Liang and Bie, Zhi and Sun, Yifan and Wang, Jingdong and Su, Chi and Wang, Shengjin and Tian, Qi},
booktitle = {European Conference on Computer Vision},
year = {2016},
pages = {868-884},
doi = {10.1007/978-3-319-46466-4_52},
url = {https://mlanthology.org/eccv/2016/zheng2016eccv-mars/}
}