Robust and Accurate Object Detection via Adversarial Learning
Abstract
Data augmentation has become a de facto component for training high-performance deep image classifiers, but its potential is under-explored for object detection. Noting that most state-of-the-art object detectors benefit from fine-tuning a pre-trained classifier, we first study how the classifiers' gains from various data augmentations transfer to object detection. The results are discouraging; the gains diminish after fine-tuning in terms of either accuracy or robustness. This work instead augments the fine-tuning stage for object detectors by exploring adversarial examples, which can be viewed as a model-dependent data augmentation. Our method dynamically selects the stronger adversarial images sourced from a detector's classification and localization branches and evolves with the detector to ensure the augmentation policy stays current and relevant. This model-dependent augmentation generalizes to different object detectors better than AutoAugment, a model-agnostic augmentation policy searched based on one particular detector. Our approach boosts the performance of state-of-the-art EfficientDets by +1.1 mAP on the COCO object detection benchmark. It also improves the detectors' robustness against natural distortions by +3.8 mAP and against domain shift by +1.3 mAP.
Cite
Text
Chen et al. "Robust and Accurate Object Detection via Adversarial Learning." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.01635Markdown
[Chen et al. "Robust and Accurate Object Detection via Adversarial Learning." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/chen2021cvpr-robust-a/) doi:10.1109/CVPR46437.2021.01635BibTeX
@inproceedings{chen2021cvpr-robust-a,
title = {{Robust and Accurate Object Detection via Adversarial Learning}},
author = {Chen, Xiangning and Xie, Cihang and Tan, Mingxing and Zhang, Li and Hsieh, Cho-Jui and Gong, Boqing},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2021},
pages = {16622-16631},
doi = {10.1109/CVPR46437.2021.01635},
url = {https://mlanthology.org/cvpr/2021/chen2021cvpr-robust-a/}
}