Adaptive Multi-Task Learning for Few-Shot Object Detection

Abstract

The majority of few-shot object detection methods use a shared feature map for both classification and localization, despite the conflicting requirements of these two tasks. Localization needs scale and positional sensitive features, whereas classification requires features that are robust to scale and positional variations. Although few methods have recognized this challenge and attempted to address it, they may not provide a comprehensive resolution to the issue. To overcome the contradictory preferences between classification and localization in few-shot object detection, an adaptive multi-task learning method, featuring a novel precision-driven gradient balancer, is proposed. This balancer effectively mitigates the conflicts by dynamically adjusting the backward gradient ratios for both tasks. Furthermore, a knowledge distillation and classification refinement scheme based on CLIP is introduced, aiming to enhance individual tasks by leveraging the capabilities of large vision-language models. Experimental results of the proposed method consistently show improvements over strong few-shot detection baselines on benchmark datasets. https://github.com/RY-Paper/MTL-FSOD

Cite

Text

Ren et al. "Adaptive Multi-Task Learning for Few-Shot Object Detection." Proceedings of the European Conference on Computer Vision (ECCV), 2024. doi:10.1007/978-3-031-72667-5_17

Markdown

[Ren et al. "Adaptive Multi-Task Learning for Few-Shot Object Detection." Proceedings of the European Conference on Computer Vision (ECCV), 2024.](https://mlanthology.org/eccv/2024/ren2024eccv-adaptive/) doi:10.1007/978-3-031-72667-5_17

BibTeX

@inproceedings{ren2024eccv-adaptive,
  title     = {{Adaptive Multi-Task Learning for Few-Shot Object Detection}},
  author    = {Ren, Yan and Li, Yanling and Kong, Adams Wai-Kin},
  booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
  year      = {2024},
  doi       = {10.1007/978-3-031-72667-5_17},
  url       = {https://mlanthology.org/eccv/2024/ren2024eccv-adaptive/}
}