Breaking the Likelihood--Quality Trade-Off in Diffusion Models by Merging Pretrained Experts

Abstract

Diffusion models have recently emerged as powerful generative models capable of producing highly realistic images. Despite their success, a persistent challenge remains: models that generate high-quality samples often assign poor likelihoods to data, and vice versa. This trade-off arises because perceptual quality depends more on modeling high-noise regions, while likelihood is dominated by sensitivity to low-level image statistics. In this work, we propose a simple yet effective method to overcome this trade-off by merging two pretrained diffusion experts, one focused on perceptual quality and the other on likelihood, within a Mixture-of-Experts framework. Our approach applies the image-quality expert during high noise levels and uses the likelihood expert in low noise levels. Empirically, our merged model consistently improves over both experts: on CIFAR-10, it achieves better likelihood and sample quality than either baseline. On ImageNet32, it matches the likelihood of the likelihood expert while surpassing the image-quality expert in FID, effectively breaking the likelihood–quality trade-off in diffusion models.

Cite

Text

Esfandiari et al. "Breaking the Likelihood--Quality Trade-Off in Diffusion Models by Merging Pretrained Experts." ICLR 2025 Workshops: DeLTa, 2025.

Markdown

[Esfandiari et al. "Breaking the Likelihood--Quality Trade-Off in Diffusion Models by Merging Pretrained Experts." ICLR 2025 Workshops: DeLTa, 2025.](https://mlanthology.org/iclrw/2025/esfandiari2025iclrw-breaking/)

BibTeX

@inproceedings{esfandiari2025iclrw-breaking,
  title     = {{Breaking the Likelihood--Quality Trade-Off in Diffusion Models by Merging Pretrained Experts}},
  author    = {Esfandiari, Yasin and Bauer, Stefan and Stich, Sebastian U and Dittadi, Andrea},
  booktitle = {ICLR 2025 Workshops: DeLTa},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/esfandiari2025iclrw-breaking/}
}