FDN: Feature Decoupling Network for Head Pose Estimation

Abstract

Head pose estimation from RGB images without depth information is a challenging task due to the loss of spatial information as well as large head pose variations in the wild. The performance of existing landmark-free methods remains unsatisfactory as the quality of estimated pose is inferior. In this paper, we propose a novel three-branch network architecture, termed as Feature Decoupling Network (FDN), a more powerful architecture for landmark-free head pose estimation from a single RGB image. In FDN, we first propose a feature decoupling (FD) module to explicitly learn the discriminative features for each pose angle by adaptively recalibrating its channel-wise responses. Besides, we introduce a cross-category center (CCC) loss to constrain the distribution of the latent variable subspaces and thus we can obtain more compact and distinct subspaces. Extensive experiments on both in-the-wild and controlled environment datasets demonstrate that the proposed method outperforms other state-of-the-art methods based on a single RGB image and behaves on par with approaches based on multimodal input resources.

Cite

Text

Zhang et al. "FDN: Feature Decoupling Network for Head Pose Estimation." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I07.6974

Markdown

[Zhang et al. "FDN: Feature Decoupling Network for Head Pose Estimation." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/zhang2020aaai-fdn/) doi:10.1609/AAAI.V34I07.6974

BibTeX

@inproceedings{zhang2020aaai-fdn,
  title     = {{FDN: Feature Decoupling Network for Head Pose Estimation}},
  author    = {Zhang, Hao and Wang, Mengmeng and Liu, Yong and Yuan, Yi},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {12789-12796},
  doi       = {10.1609/AAAI.V34I07.6974},
  url       = {https://mlanthology.org/aaai/2020/zhang2020aaai-fdn/}
}