Hi4D: 4D Instance Segmentation of Close Human Interaction
Abstract
We propose Hi4D, a method and dataset for the auto analysis of physically close human-human interaction under prolonged contact. Robustly disentangling several in-contact subjects is a challenging task due to occlusions and complex shapes. Hence, existing multi-view systems typically fuse 3D surfaces of close subjects into a single, connected mesh. To address this issue we leverage i) individually fitted neural implicit avatars; ii) an alternating optimization scheme that refines pose and surface through periods of close proximity; and iii) thus segment the fused raw scans into individual instances. From these instances we compile Hi4D dataset of 4D textured scans of 20 subject pairs, 100 sequences, and a total of more than 11K frames. Hi4D contains rich interaction-centric annotations in 2D and 3D alongside accurately registered parametric body models. We define varied human pose and shape estimation tasks on this dataset and provide results from state-of-the-art methods on these benchmarks.
Cite
Text
Yin et al. "Hi4D: 4D Instance Segmentation of Close Human Interaction." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.01632Markdown
[Yin et al. "Hi4D: 4D Instance Segmentation of Close Human Interaction." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/yin2023cvpr-hi4d/) doi:10.1109/CVPR52729.2023.01632BibTeX
@inproceedings{yin2023cvpr-hi4d,
title = {{Hi4D: 4D Instance Segmentation of Close Human Interaction}},
author = {Yin, Yifei and Guo, Chen and Kaufmann, Manuel and Zarate, Juan Jose and Song, Jie and Hilliges, Otmar},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2023},
pages = {17016-17027},
doi = {10.1109/CVPR52729.2023.01632},
url = {https://mlanthology.org/cvpr/2023/yin2023cvpr-hi4d/}
}