Deep Virtual Markers for Articulated 3D Shapes

Abstract

We propose deep virtual markers, a framework for estimating dense and accurate positional information for various types of 3D data. We design a concept and construct a framework that maps 3D points of 3D articulated models, like humans, into virtual marker labels. To realize the framework, we adopt a sparse convolutional neural network and classify 3D points of an articulated model into virtual marker labels. We propose to use soft labels for the classifier to learn rich and dense interclass relationships based on geodesic distance. To measure the localization accuracy of the virtual markers, we test FAUST challenge, and our result outperforms the state-of-the-art. We also observe outstanding performance on the generalizability test, unseen data evaluation, and different 3D data types (meshes and depth maps). We show additional applications using the estimated virtual markers, such as non-rigid registration, texture transfer, and realtime dense marker prediction from depth maps.

Cite

Text

Kim et al. "Deep Virtual Markers for Articulated 3D Shapes." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.01141

Markdown

[Kim et al. "Deep Virtual Markers for Articulated 3D Shapes." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/kim2021iccv-deep/) doi:10.1109/ICCV48922.2021.01141

BibTeX

@inproceedings{kim2021iccv-deep,
  title     = {{Deep Virtual Markers for Articulated 3D Shapes}},
  author    = {Kim, Hyomin and Kim, Jungeon and Kam, Jaewon and Park, Jaesik and Lee, Seungyong},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {11615-11625},
  doi       = {10.1109/ICCV48922.2021.01141},
  url       = {https://mlanthology.org/iccv/2021/kim2021iccv-deep/}
}