IsMo-GAN: Adversarial Learning for Monocular Non-Rigid 3D Reconstruction
Abstract
The majority of the existing methods for non-rigid 3D surface regression from a single 2D image require an object template or point tracks over multiple frames as an input, and are still far from real-time processing rates. In this work, we present the Isometry-Aware Monocular Generative Adversarial Network (IsMo-GAN) — an approach for direct 3D reconstruction from a single image, trained for the deformation model in an adversarial manner on a light-weight synthetic dataset. IsMo-GAN reconstructs surfaces from real images under varying illumination, camera poses, textures and shading at over 250 Hz. In multiple experiments, it consistently outperforms multiple approaches in the reconstruction accuracy, runtime, generalisation to unknown surfaces and robustness to occlusions. In comparison to the state-of-the-art, we reduce the reconstruction error by 10-30% including the textureless case and our surfaces evince fewer artefacts qualitatively.
Cite
Text
Shimada et al. "IsMo-GAN: Adversarial Learning for Monocular Non-Rigid 3D Reconstruction." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019. doi:10.1109/CVPRW.2019.00347Markdown
[Shimada et al. "IsMo-GAN: Adversarial Learning for Monocular Non-Rigid 3D Reconstruction." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/shimada2019cvprw-ismogan/) doi:10.1109/CVPRW.2019.00347BibTeX
@inproceedings{shimada2019cvprw-ismogan,
title = {{IsMo-GAN: Adversarial Learning for Monocular Non-Rigid 3D Reconstruction}},
author = {Shimada, Soshi and Golyanik, Vladislav and Theobalt, Christian and Stricker, Didier},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2019},
pages = {2876-2885},
doi = {10.1109/CVPRW.2019.00347},
url = {https://mlanthology.org/cvprw/2019/shimada2019cvprw-ismogan/}
}