Learning to Quantize for Training Vector-Quantized Networks

Abstract

Deep neural networks incorporating discrete latent variables have shown significant potential in sequence modeling. A notable approach is to leverage vector quantization (VQ) to generate discrete representations within a codebook. However, its discrete nature prevents the use of standard backpropagation, which has led to challenges in efficient codebook training. In this work, we introduce Meta-Quantization (MQ), a novel vector quantization training framework inspired by meta-learning. Our method separates the optimization of the codebook and the auto-encoder into two levels. Furthermore, we introduce a hyper-net to replace the embedding-parameterized codebook, enabling the codebook to be dynamically generated based on the feedback from the auto-encoder. Different from previous VQ objectives, our innovation results in a meta-objective that makes the codebook training task-aware. We validate the effectiveness of MQ with VQVAE and VQGAN architecture on image reconstruction and generation tasks. Experimental results showcase the superior generative performance of MQ, underscoring its potential as a robust alternative to existing VQ methods.

Cite

Text

Qin and Zhang. "Learning to Quantize for Training Vector-Quantized Networks." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Qin and Zhang. "Learning to Quantize for Training Vector-Quantized Networks." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/qin2025icml-learning/)

BibTeX

@inproceedings{qin2025icml-learning,
  title     = {{Learning to Quantize for Training Vector-Quantized Networks}},
  author    = {Qin, Peijia and Zhang, Jianguo},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {50435-50447},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/qin2025icml-learning/}
}