PG-Net: Pixel to Global Matching Network for Visual Tracking

Abstract

Siamese neural network has been well investigated by tracking frameworks due to its fast speed and high accuracy. However, very few efforts were spent on background-extraction by those approaches. In this paper, a Pixel to Global Matching Network (PG-Net) is proposed to suppress the influence of background in search image while achieving state-of-the-art tracking performance. To achieve this purpose, each pixel on search feature is utilized to calculate the similarity with global template feature. This calculation method can appropriately reduce the matching area, thus introducing less background interference. In addition, we propose a new tracking framework to perform correlation-shared tracking and multiple losses for training, which not only reduce the computational burden but also improve the performance. We conduct comparison experiments on various public tracking datasets, which obtains state-of-the-art performance while running with fast speed.

Cite

Text

Liao et al. "PG-Net: Pixel to Global Matching Network for Visual Tracking." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58542-6_26

Markdown

[Liao et al. "PG-Net: Pixel to Global Matching Network for Visual Tracking." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/liao2020eccv-pgnet/) doi:10.1007/978-3-030-58542-6_26

BibTeX

@inproceedings{liao2020eccv-pgnet,
  title     = {{PG-Net: Pixel to Global Matching Network for Visual Tracking}},
  author    = {Liao, Bingyan and Wang, Chenye and Wang, Yayun and Wang, Yaonong and Yin, Jun},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2020},
  doi       = {10.1007/978-3-030-58542-6_26},
  url       = {https://mlanthology.org/eccv/2020/liao2020eccv-pgnet/}
}