Leveraging Conditional Dependence for Efficient World Model Denoising

Abstract

Effective denoising is critical for managing complex visual inputs contaminated with noisy distractors in model-based reinforcement learning (RL). Current methods often oversimplify the decomposition of observations by neglecting the conditional dependence between task-relevant and task-irrelevant components given an observation. To address this limitation, we introduce CsDreamer, a model-based RL approach built upon the world model of Collider-structure Recurrent State-Space Model (CsRSSM). CsRSSM incorporates colliders to comprehensively model the denoising inference process and explicitly capture the conditional dependence. Furthermore, it employs a decoupling regularization to balance the influence of this conditional dependence. By accurately inferring a task-relevant state space, CsDreamer improves learning efficiency during rollouts. Experimental results demonstrate the effectiveness of CsRSSM in extracting task-relevant information, leading to CsDreamer outperforming existing approaches in environments characterized by complex noise interference.

Cite

Text

Zhang et al. "Leveraging Conditional Dependence for Efficient World Model Denoising." Advances in Neural Information Processing Systems, 2025.

Markdown

[Zhang et al. "Leveraging Conditional Dependence for Efficient World Model Denoising." Advances in Neural Information Processing Systems, 2025.](https://mlanthology.org/neurips/2025/zhang2025neurips-leveraging/)

BibTeX

@inproceedings{zhang2025neurips-leveraging,
  title     = {{Leveraging Conditional Dependence for Efficient World Model Denoising}},
  author    = {Zhang, Shaowei and Cao, Jiahan and Cheng, Dian and Zhou, Xunlan and Wan, Shenghua and Gan, Le and Zhan, De-Chuan},
  booktitle = {Advances in Neural Information Processing Systems},
  year      = {2025},
  url       = {https://mlanthology.org/neurips/2025/zhang2025neurips-leveraging/}
}