Du²Net: Learning Depth Estimation from Dual-Cameras and Dual-Pixels
Abstract
Computational stereo has reached a high level of accuracy, but degrades in the presence of occlusions, repeated textures, and correspondence errors along edges. We present a novel approach based on neural networks for depth estimation that combines stereo from dual cameras with stereo from a dual-pixel sensor, which is increasingly common on consumer cameras. Our network uses a novel architecture to fuse these two sources of information and can overcome the above-mentioned limitations of pure binocular stereo matching. Our method provides a dense depth map with sharp edges, which is crucial for computational photography applications like synthethic shallow-depth-of-field or 3D Photos. Additionally, we avoid the inherent ambiguity due to the aperture problem in stereo cameras by designing the stereo baseline to be orthogonal to the dual-pixel baseline. We present experiments and comparisons with state-of-the-art approaches that show that our method offers a substantial improvement over previous works.
Cite
Text
Zhang et al. "Du²Net: Learning Depth Estimation from Dual-Cameras and Dual-Pixels." Proceedings of the European Conference on Computer Vision (ECCV), 2020. doi:10.1007/978-3-030-58452-8_34Markdown
[Zhang et al. "Du²Net: Learning Depth Estimation from Dual-Cameras and Dual-Pixels." Proceedings of the European Conference on Computer Vision (ECCV), 2020.](https://mlanthology.org/eccv/2020/zhang2020eccv-du2net/) doi:10.1007/978-3-030-58452-8_34BibTeX
@inproceedings{zhang2020eccv-du2net,
title = {{Du²Net: Learning Depth Estimation from Dual-Cameras and Dual-Pixels}},
author = {Zhang, Yinda and Wadhwa, Neal and Orts-Escolano, Sergio and Häne, Christian and Fanello, Sean and Garg, Rahul},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2020},
doi = {10.1007/978-3-030-58452-8_34},
url = {https://mlanthology.org/eccv/2020/zhang2020eccv-du2net/}
}