Lossy Image Compression with Quantized Hierarchical VAEs

Abstract

Recent work has shown a strong theoretical connection between variational autoencoders (VAEs) and the rate distortion theory. Motivated by this, we consider the problem of lossy image compression from the perspective of generative modeling. Starting from ResNet VAEs, which are originally designed for data (image) distribution modeling, we redesign their latent variable model using a quantization-aware posterior and prior, enabling easy quantization and entropy coding for image compression. Along with improved neural network blocks, we present a powerful and efficient class of lossy image coders, outperforming previous methods on natural image (lossy) compression. Our model compresses images in a coarse-to-fine fashion and supports parallel encoding and decoding, leading to fast execution on GPUs.

Cite

Text

Duan et al. "Lossy Image Compression with Quantized Hierarchical VAEs." Winter Conference on Applications of Computer Vision, 2023.

Markdown

[Duan et al. "Lossy Image Compression with Quantized Hierarchical VAEs." Winter Conference on Applications of Computer Vision, 2023.](https://mlanthology.org/wacv/2023/duan2023wacv-lossy/)

BibTeX

@inproceedings{duan2023wacv-lossy,
  title     = {{Lossy Image Compression with Quantized Hierarchical VAEs}},
  author    = {Duan, Zhihao and Lu, Ming and Ma, Zhan and Zhu, Fengqing},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2023},
  pages     = {198-207},
  url       = {https://mlanthology.org/wacv/2023/duan2023wacv-lossy/}
}