Deep Depth from Focus with Differential Focus Volume

Abstract

Depth-from-focus (DFF) is a technique that infers depth using the focus change of a camera. In this work, we propose a convolutional neural network (CNN) to find the best-focused pixels in a focal stack and infer depth from the focus estimation. The key innovation of the network is the novel deep differential focus volume (DFV). By computing the first-order derivative with the stacked features over different focal distances, DFV is able to capture both the focus and context information for focus analysis. Besides, we also introduce a probability regression mechanism for focus estimation to handle sparsely sampled focal stacks and provide uncertainty estimation to the final prediction. Comprehensive experiments demonstrate that the proposed model achieves state-of-the-art performance on multiple datasets with good generalizability and fast speed.

Cite

Text

Yang et al. "Deep Depth from Focus with Differential Focus Volume." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.01231

Markdown

[Yang et al. "Deep Depth from Focus with Differential Focus Volume." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/yang2022cvpr-deep/) doi:10.1109/CVPR52688.2022.01231

BibTeX

@inproceedings{yang2022cvpr-deep,
  title     = {{Deep Depth from Focus with Differential Focus Volume}},
  author    = {Yang, Fengting and Huang, Xiaolei and Zhou, Zihan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {12642-12651},
  doi       = {10.1109/CVPR52688.2022.01231},
  url       = {https://mlanthology.org/cvpr/2022/yang2022cvpr-deep/}
}