LayoutDM: Discrete Diffusion Model for Controllable Layout Generation

Abstract

Controllable layout generation aims at synthesizing plausible arrangement of element bounding boxes with optional constraints, such as type or position of a specific element. In this work, we try to solve a broad range of layout generation tasks in a single model that is based on discrete state-space diffusion models. Our model, named LayoutDM, naturally handles the structured layout data in the discrete representation and learns to progressively infer a noiseless layout from the initial input, where we model the layout corruption process by modality-wise discrete diffusion. For conditional generation, we propose to inject layout constraints in the form of masking or logit adjustment during inference. We show in the experiments that our LayoutDM successfully generates high-quality layouts and outperforms both task-specific and task-agnostic baselines on several layout tasks.

Cite

Text

Inoue et al. "LayoutDM: Discrete Diffusion Model for Controllable Layout Generation." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00980

Markdown

[Inoue et al. "LayoutDM: Discrete Diffusion Model for Controllable Layout Generation." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/inoue2023cvpr-layoutdm/) doi:10.1109/CVPR52729.2023.00980

BibTeX

@inproceedings{inoue2023cvpr-layoutdm,
  title     = {{LayoutDM: Discrete Diffusion Model for Controllable Layout Generation}},
  author    = {Inoue, Naoto and Kikuchi, Kotaro and Simo-Serra, Edgar and Otani, Mayu and Yamaguchi, Kota},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {10167-10176},
  doi       = {10.1109/CVPR52729.2023.00980},
  url       = {https://mlanthology.org/cvpr/2023/inoue2023cvpr-layoutdm/}
}