LD-RPS: Zero-Shot Unified Image Restoration via Latent Diffusion Recurrent Posterior Sampling

Abstract

Unified image restoration is a significantly challenging task in low-level vision. Existing methods either make tailored designs for specific tasks, limiting their generalizability across various types of degradation, or rely on training with paired datasets, thereby suffering from closed-set constraints. To address these issues, we propose a novel, dataset-free, and unified approach through recurrent posterior sampling utilizing a pretrained latent diffusion model. Our method incorporates the multimodal understanding model to provide sematic priors for the generative model under a task-blind condition. Furthermore, it utilizes a lightweight module to align the degraded input with the generated preference of the diffusion model, and employs recurrent refinement for posterior sampling. Extensive experiments demonstrate that our method outperforms state-of-the-art methods, validating its effectiveness and robustness. Our code and data will be available at https://github.com/AMAP-ML/LD-RPS.

Cite

Text

Li et al. "LD-RPS: Zero-Shot Unified Image Restoration via Latent Diffusion Recurrent Posterior Sampling." International Conference on Computer Vision, 2025.

Markdown

[Li et al. "LD-RPS: Zero-Shot Unified Image Restoration via Latent Diffusion Recurrent Posterior Sampling." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/li2025iccv-ldrps/)

BibTeX

@inproceedings{li2025iccv-ldrps,
  title     = {{LD-RPS: Zero-Shot Unified Image Restoration via Latent Diffusion Recurrent Posterior Sampling}},
  author    = {Li, Huaqiu and Wang, Yong and Huang, Tongwen and Huang, Hailang and Wang, Haoqian and Chu, Xiangxiang},
  booktitle = {International Conference on Computer Vision},
  year      = {2025},
  pages     = {13684-13694},
  url       = {https://mlanthology.org/iccv/2025/li2025iccv-ldrps/}
}