Deep Image Set Hashing
Abstract
In applications involving matching of image sets, the information from multiple images must be effectively exploited to represent each set. State-of-the-art methods use probabilistic distribution or subspace to model a set and use specific distance measure to compare two sets. These methods are slow to compute and not compact to use in a large scale scenario. Learning-based hashing is often used in large scale image retrieval as they provide a compact representation of each sample and the Hamming distance can be used to efficiently compare two samples. However, most hashing methods encode each image separately and discard knowledge that multiple images in the same set represent the same object or person. We investigate the set hashing problem by combining both set representation and hashing in a single deep neural network. An image set is first passed to a CNN module to extract image features, then these features are aggregated using two types of set feature to capture both set specific and database-wide distribution information. The computed set feature is then fed into a multilayer perceptron to learn a compact binary embedding trained with triplet loss. We extensively evaluate our approach on multiple image datasets and show highly competitive performance compared to state-of-the-art methods.
Cite
Text
Feng et al. "Deep Image Set Hashing." IEEE/CVF Winter Conference on Applications of Computer Vision, 2017. doi:10.1109/WACV.2017.143Markdown
[Feng et al. "Deep Image Set Hashing." IEEE/CVF Winter Conference on Applications of Computer Vision, 2017.](https://mlanthology.org/wacv/2017/feng2017wacv-deep/) doi:10.1109/WACV.2017.143BibTeX
@inproceedings{feng2017wacv-deep,
title = {{Deep Image Set Hashing}},
author = {Feng, Jie and Karaman, Svebor and Chang, Shih-Fu},
booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
year = {2017},
pages = {1241-1250},
doi = {10.1109/WACV.2017.143},
url = {https://mlanthology.org/wacv/2017/feng2017wacv-deep/}
}