Distribution-Aligned Diffusion for Human Mesh Recovery

Abstract

Recovering a 3D human mesh from a single RGB image is a challenging task due to depth ambiguity and self-occlusion, resulting in a high degree of uncertainty. Meanwhile, diffusion models have recently seen much success in generating high-quality outputs by progressively denoising noisy inputs. Inspired by their capability, we explore a diffusion-based approach for human mesh recovery, and propose a Human Mesh Diffusion (HMDiff) framework which frames mesh recovery as a reverse diffusion process. We also propose a Distribution Alignment Technique (DAT) that injects input-specific distribution information into the diffusion process, and provides useful prior knowledge to simplify the mesh recovery task. Our method achieves state-of-the-art performance on three widely used datasets. Project page: https://gongjia0208.github.io/HMDiff/.

Cite

Text

Foo et al. "Distribution-Aligned Diffusion for Human Mesh Recovery." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.00846

Markdown

[Foo et al. "Distribution-Aligned Diffusion for Human Mesh Recovery." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/foo2023iccv-distributionaligned/) doi:10.1109/ICCV51070.2023.00846

BibTeX

@inproceedings{foo2023iccv-distributionaligned,
  title     = {{Distribution-Aligned Diffusion for Human Mesh Recovery}},
  author    = {Foo, Lin Geng and Gong, Jia and Rahmani, Hossein and Liu, Jun},
  booktitle = {International Conference on Computer Vision},
  year      = {2023},
  pages     = {9221-9232},
  doi       = {10.1109/ICCV51070.2023.00846},
  url       = {https://mlanthology.org/iccv/2023/foo2023iccv-distributionaligned/}
}