Conditional Hallucinations for Image Compression
Abstract
In lossy image compression, models face the challenge of either hallucinating details or generating out-of-distribution samples due to the information bottleneck. This implies that at times, introducing hallucinations is necessary to generate in-distribution samples. The optimal level of hallucination varies depending on image content, as humans are sensitive to small changes that alter the semantic meaning. We propose a novel compression method that dynamically balances the degree of hallucination based on content. We collect data and train a model to predict user preferences on hallucinations. By using this prediction to adjust the perceptual weight in the reconstruction loss, we develop a \textbf{Con}ditionally \textbf{Ha}llucinating compression model (\textbf{ConHa}) that outperforms state-of-the-art image compression methods. Code and images are available at \href{https://polybox.ethz.ch/index.php/s/owS1k5JYs4KD4TA}https://polybox.ethz.ch/index.php/s/owS1k5JYs4KD4TA.
Cite
Text
Aczel and Wattenhofer. "Conditional Hallucinations for Image Compression." NeurIPS 2024 Workshops: Compression, 2024.Markdown
[Aczel and Wattenhofer. "Conditional Hallucinations for Image Compression." NeurIPS 2024 Workshops: Compression, 2024.](https://mlanthology.org/neuripsw/2024/aczel2024neuripsw-conditional/)BibTeX
@inproceedings{aczel2024neuripsw-conditional,
title = {{Conditional Hallucinations for Image Compression}},
author = {Aczel, Till and Wattenhofer, Roger},
booktitle = {NeurIPS 2024 Workshops: Compression},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/aczel2024neuripsw-conditional/}
}