Hand-Object Interaction Image Generation
Abstract
In this work, we are dedicated to a new task, i.e., hand-object interaction image generation, which aims to conditionally generate the hand-object image under the given hand, object and their interaction status. This task is challenging and research-worthy in many potential application scenarios, such as AR/VR games and online shopping, etc. To address this problem, we propose a novel HOGAN framework, which utilizes the expressive model-aware hand-object representation and leverages its inherent topology to build the unified surface space. In this space, we explicitly consider the complex self- and mutual occlusion during interaction. During final image synthesis, we consider different characteristics of hand and object and generate the target image in a split-and-combine manner. For evaluation, we build a comprehensive protocol to access both the fidelity and structure preservation of the generated image. Extensive experiments on two large-scale datasets, i.e., HO3Dv3 and DexYCB, demonstrate the effectiveness and superiority of our framework both quantitatively and qualitatively. The code will be available at https://github.com/play-with-HOI-generation/HOIG.
Cite
Text
Hu et al. "Hand-Object Interaction Image Generation." Neural Information Processing Systems, 2022.Markdown
[Hu et al. "Hand-Object Interaction Image Generation." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/hu2022neurips-handobject/)BibTeX
@inproceedings{hu2022neurips-handobject,
title = {{Hand-Object Interaction Image Generation}},
author = {Hu, Hezhen and Wang, Weilun and Zhou, Wengang and Li, Houqiang},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/hu2022neurips-handobject/}
}