Ultra Lowrate Image Compression with Semantic Residual Coding and Compression-Aware Diffusion

Abstract

Existing multimodal large model-based image compression frameworks often rely on a fragmented integration of semantic retrieval, latent compression, and generative models, resulting in suboptimal performance in both reconstruction fidelity and coding efficiency. To address these challenges, we propose a residual-guided ultra lowrate image compression named ResULIC, which incorporates residual signals into both semantic retrieval and the diffusion-based generation process. Specifically, we introduce Semantic Residual Coding (SRC) to capture the semantic disparity between the original image and its compressed latent representation. A perceptual fidelity optimizer is further applied for superior reconstruction quality. Additionally, we present the Compression-aware Diffusion Model (CDM), which establishes an optimal alignment between bitrates and diffusion time steps, improving compression-reconstruction synergy. Extensive experiments demonstrate the effectiveness of ResULIC, achieving superior objective and subjective performance compared to state-of-the-art diffusion-based methods with -80.7%, -66.3% BD-rate saving in terms of LPIPS and FID.

Cite

Text

Ke et al. "Ultra Lowrate Image Compression with Semantic Residual Coding and Compression-Aware Diffusion." Proceedings of the 42nd International Conference on Machine Learning, 2025.

Markdown

[Ke et al. "Ultra Lowrate Image Compression with Semantic Residual Coding and Compression-Aware Diffusion." Proceedings of the 42nd International Conference on Machine Learning, 2025.](https://mlanthology.org/icml/2025/ke2025icml-ultra/)

BibTeX

@inproceedings{ke2025icml-ultra,
  title     = {{Ultra Lowrate Image Compression with Semantic Residual Coding and Compression-Aware Diffusion}},
  author    = {Ke, Anle and Zhang, Xu and Chen, Tong and Lu, Ming and Zhou, Chao and Gu, Jiawen and Ma, Zhan},
  booktitle = {Proceedings of the 42nd International Conference on Machine Learning},
  year      = {2025},
  pages     = {29626-29650},
  volume    = {267},
  url       = {https://mlanthology.org/icml/2025/ke2025icml-ultra/}
}