Semi-Automatic Annotation for Visual Object Tracking

Abstract

We propose a semi-automatic bounding box annotation method for visual object tracking by utilizing temporal information with a tracking-by-detection approach. For detection, we use an off-the-shelf object detector which is trained iteratively with the annotations generated by the proposed method, and we perform object detection on each frame independently. We employ Multiple Hypothesis Tracking (MHT) to exploit temporal information and to reduce the number of false-positives which makes it possible to use lower objectness thresholds for detection to increase recall. The tracklets formed by MHT are evaluated by human operators to enlarge the training set. This novel incremental learning approach helps to perform annotation iteratively. The experiments performed on AUTH Multidrone Dataset reveal that the annotation workload can be reduced up to 96% by the proposed approach. Resulting uav_detection_2 annotations and our codes are publicly available at github.com/aybora/Semi-Automatic-Video-Annotation-OGAM.

Cite

Text

Ince et al. "Semi-Automatic Annotation for Visual Object Tracking." IEEE/CVF International Conference on Computer Vision Workshops, 2021. doi:10.1109/ICCVW54120.2021.00143

Markdown

[Ince et al. "Semi-Automatic Annotation for Visual Object Tracking." IEEE/CVF International Conference on Computer Vision Workshops, 2021.](https://mlanthology.org/iccvw/2021/ince2021iccvw-semiautomatic/) doi:10.1109/ICCVW54120.2021.00143

BibTeX

@inproceedings{ince2021iccvw-semiautomatic,
  title     = {{Semi-Automatic Annotation for Visual Object Tracking}},
  author    = {Ince, Kutalmis Gokalp and Koksal, Aybora and Fazla, Arda and Alatan, A. Aydin},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2021},
  pages     = {1233-1239},
  doi       = {10.1109/ICCVW54120.2021.00143},
  url       = {https://mlanthology.org/iccvw/2021/ince2021iccvw-semiautomatic/}
}