Learning Efficient Photometric Feature Transform for Multi-View Stereo
Abstract
We present a novel framework to learn to convert the per-pixel photometric information at each view into spatially distinctive and view-invariant low-level features, which can be plugged into existing multi-view stereo pipeline for enhanced 3D reconstruction. Both the illumination conditions during acquisition and the subsequent per-pixel feature transform can be jointly optimized in a differentiable fashion. Our framework automatically adapts to and makes efficient use of the geometric information available in different forms of input data. High-quality 3D reconstructions of a variety of challenging objects are demonstrated on the data captured with an illumination multiplexing device, as well as a point light. Our results compare favorably with state-of-the-art techniques.
Cite
Text
Kang et al. "Learning Efficient Photometric Feature Transform for Multi-View Stereo." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.00590Markdown
[Kang et al. "Learning Efficient Photometric Feature Transform for Multi-View Stereo." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/kang2021iccv-learning/) doi:10.1109/ICCV48922.2021.00590BibTeX
@inproceedings{kang2021iccv-learning,
title = {{Learning Efficient Photometric Feature Transform for Multi-View Stereo}},
author = {Kang, Kaizhang and Xie, Cihui and Zhu, Ruisheng and Ma, Xiaohe and Tan, Ping and Wu, Hongzhi and Zhou, Kun},
booktitle = {International Conference on Computer Vision},
year = {2021},
pages = {5956-5965},
doi = {10.1109/ICCV48922.2021.00590},
url = {https://mlanthology.org/iccv/2021/kang2021iccv-learning/}
}