SurfaceNet: An End-to-End 3D Neural Network for Multiview Stereopsis
Abstract
This paper proposes an end-to-end learning framework for multiview stereopsis. We term the network SurfaceNet. It takes a set of images and their corresponding camera parameters as input and directly infers the 3D model. The key advantage of the framework is that both photo-consistency as well geometric relations of the surface structure can be directly learned for the purpose of multiview stereopsis in an end-to-end fashion. SurfaceNet is a fully 3D convolutional network which is achieved by encoding the camera parameters together with the images in a 3D voxel representation. We evaluate SurfaceNet on the large-scale DTU benchmark.
Cite
Text
Ji et al. "SurfaceNet: An End-to-End 3D Neural Network for Multiview Stereopsis." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.253Markdown
[Ji et al. "SurfaceNet: An End-to-End 3D Neural Network for Multiview Stereopsis." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/ji2017iccv-surfacenet/) doi:10.1109/ICCV.2017.253BibTeX
@inproceedings{ji2017iccv-surfacenet,
title = {{SurfaceNet: An End-to-End 3D Neural Network for Multiview Stereopsis}},
author = {Ji, Mengqi and Gall, Juergen and Zheng, Haitian and Liu, Yebin and Fang, Lu},
booktitle = {International Conference on Computer Vision},
year = {2017},
doi = {10.1109/ICCV.2017.253},
url = {https://mlanthology.org/iccv/2017/ji2017iccv-surfacenet/}
}