Sparse Quantization for Patch Description

Abstract

The representation of local image patches is crucial for the good performance and efficiency of many vision tasks. Patch descriptors have been designed to generalize towards diverse variations, depending on the application, as well as the desired compromise between accuracy and efficiency. We present a novel formulation of patch description, that serves such issues well. Sparse quantization lies at its heart. This allows for efficient encodings, leading to powerful, novel binary descriptors, yet also to the generalization of existing descriptors like SIFT or BRIEF. We demonstrate the capabilities of our formulation for both keypoint matching and image classification. Our binary descriptors achieve state-of-the-art results for two keypoint matching benchmarks, namely those by Brown [6] and Mikolajczyk [18]. For image classification, we propose new descriptors that perform similar to SIFT on Caltech101 [10] and PASCAL VOC07 [9].

Cite

Text

Boix et al. "Sparse Quantization for Patch Description." Conference on Computer Vision and Pattern Recognition, 2013. doi:10.1109/CVPR.2013.366

Markdown

[Boix et al. "Sparse Quantization for Patch Description." Conference on Computer Vision and Pattern Recognition, 2013.](https://mlanthology.org/cvpr/2013/boix2013cvpr-sparse/) doi:10.1109/CVPR.2013.366

BibTeX

@inproceedings{boix2013cvpr-sparse,
  title     = {{Sparse Quantization for Patch Description}},
  author    = {Boix, Xavier and Gygli, Michael and Roig, Gemma and Van Gool, Luc},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2013},
  doi       = {10.1109/CVPR.2013.366},
  url       = {https://mlanthology.org/cvpr/2013/boix2013cvpr-sparse/}
}