Shadow Generation for Composite Image Using Diffusion Model

Abstract

In the realm of image composition generating realistic shadow for the inserted foreground remains a formidable challenge. Previous works have developed image-to-image translation models which are trained on paired training data. However they are struggling to generate shadows with accurate shapes and intensities hindered by data scarcity and inherent task complexity. In this paper we resort to foundation model with rich prior knowledge of natural shadow images. Specifically we first adapt ControlNet to our task and then propose intensity modulation modules to improve the shadow intensity. Moreover we extend the small-scale DESOBA dataset to DESOBAv2 using a novel data acquisition pipeline. Experimental results on both DESOBA and DESOBAv2 datasets as well as real composite images demonstrate the superior capability of our model for shadow generation task. The dataset code and model are released at https://github.com/bcmi/Object-Shadow-Generation-Dataset-DESOBAv2.

Cite

Text

Liu et al. "Shadow Generation for Composite Image Using Diffusion Model." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.00776

Markdown

[Liu et al. "Shadow Generation for Composite Image Using Diffusion Model." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/liu2024cvpr-shadow/) doi:10.1109/CVPR52733.2024.00776

BibTeX

@inproceedings{liu2024cvpr-shadow,
  title     = {{Shadow Generation for Composite Image Using Diffusion Model}},
  author    = {Liu, Qingyang and You, Junqi and Wang, Jianting and Tao, Xinhao and Zhang, Bo and Niu, Li},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {8121-8130},
  doi       = {10.1109/CVPR52733.2024.00776},
  url       = {https://mlanthology.org/cvpr/2024/liu2024cvpr-shadow/}
}