Vector Quantization with Sorting Transformation

Abstract

Vector quantization is a compression technique for vector data. It creates a collection of codewords to represent the entire vector space. Each vector data is then represented by its nearest neighbor codeword, where the distance between them is the compression error. To improve nearest neighbor representation for vector quantization, we propose to apply sorting transformation to vector data. It can be shown that among all permutation transformations, the sorting transformation minimizes L2 distance and maximizes similarity measures such as cosine similarity and Pearson correlation for vector data. Through experimental validation, we show that sorting transformation based vector quantization prominently reduces compression errors and improves nearest neighbor retrieval performance.

Cite

Text

Wang and Syeda-mahmood. "Vector Quantization with Sorting Transformation." NeurIPS 2024 Workshops: Compression, 2024.

Markdown

[Wang and Syeda-mahmood. "Vector Quantization with Sorting Transformation." NeurIPS 2024 Workshops: Compression, 2024.](https://mlanthology.org/neuripsw/2024/wang2024neuripsw-vector/)

BibTeX

@inproceedings{wang2024neuripsw-vector,
  title     = {{Vector Quantization with Sorting Transformation}},
  author    = {Wang, Hongzhi and Syeda-mahmood, Tanveer},
  booktitle = {NeurIPS 2024 Workshops: Compression},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/wang2024neuripsw-vector/}
}