Keypoint-Aligned Embeddings for Image Retrieval and Re-Identification
Abstract
Learning embeddings that are invariant to the pose of the object is crucial in visual image retrieval and re-identification. The existing approaches for person, vehicle, or animal re-identification tasks suffer from high intra-class variance due to deformable shapes and different camera viewpoints. To overcome this limitation, we propose to align the image embedding with a predefined order of the keypoints. The proposed keypoint aligned embeddings model (KAE-Net) learns part-level features via multi-task learning which is guided by keypoint locations. More specifically, KAE-Net extracts channels from a feature map activated by a specific keypoint through learning the auxiliary task of heatmap reconstruction for this keypoint. The KAE-Net is compact, generic and conceptually simple. It achieves state of the art performance on the benchmark datasets of CUB-200-2011, Cars196 and VeRi-776 for retrieval and re-identification tasks.
Cite
Text
Moskvyak et al. "Keypoint-Aligned Embeddings for Image Retrieval and Re-Identification." Winter Conference on Applications of Computer Vision, 2021.Markdown
[Moskvyak et al. "Keypoint-Aligned Embeddings for Image Retrieval and Re-Identification." Winter Conference on Applications of Computer Vision, 2021.](https://mlanthology.org/wacv/2021/moskvyak2021wacv-keypointaligned/)BibTeX
@inproceedings{moskvyak2021wacv-keypointaligned,
title = {{Keypoint-Aligned Embeddings for Image Retrieval and Re-Identification}},
author = {Moskvyak, Olga and Maire, Frederic and Dayoub, Feras and Baktashmotlagh, Mahsa},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2021},
pages = {676-685},
url = {https://mlanthology.org/wacv/2021/moskvyak2021wacv-keypointaligned/}
}