TransMVSNet: Global Context-Aware Multi-View Stereo Network with Transformers

Abstract

In this paper, we present TransMVSNet, based on our exploration of feature matching in multi-view stereo (MVS). We analogize MVS back to its nature of a feature matching task and therefore propose a powerful Feature Matching Transformer (FMT) to leverage intra- (self-) and inter- (cross-) attention to aggregate long-range context information within and across images. To facilitate a better adaptation of the FMT, we leverage an Adaptive Receptive Field (ARF) module to ensure a smooth transit in scopes of features and bridge different stages with a feature pathway to pass transformed features and gradients across different scales. In addition, we apply pair-wise feature correlation to measure similarity between features, and adopt ambiguity-reducing focal loss to strengthen the supervision. To the best of our knowledge, TransMVSNet is the first attempt to leverage Transformer into the task of MVS. As a result, our method achieves state-of-the-art performance on DTU dataset, Tanks and Temples benchmark and BlendedMVS dataset. Code is available at https://github.com/MegviiRobot/TransMVSNet.

Cite

Text

Ding et al. "TransMVSNet: Global Context-Aware Multi-View Stereo Network with Transformers." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00839

Markdown

[Ding et al. "TransMVSNet: Global Context-Aware Multi-View Stereo Network with Transformers." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/ding2022cvpr-transmvsnet/) doi:10.1109/CVPR52688.2022.00839

BibTeX

@inproceedings{ding2022cvpr-transmvsnet,
  title     = {{TransMVSNet: Global Context-Aware Multi-View Stereo Network with Transformers}},
  author    = {Ding, Yikang and Yuan, Wentao and Zhu, Qingtian and Zhang, Haotian and Liu, Xiangyue and Wang, Yuanjiang and Liu, Xiao},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {8585-8594},
  doi       = {10.1109/CVPR52688.2022.00839},
  url       = {https://mlanthology.org/cvpr/2022/ding2022cvpr-transmvsnet/}
}