LCD: Learned Cross-Domain Descriptors for 2D-3D Matching
Abstract
In this work, we present a novel method to learn a local cross-domain descriptor for 2D image and 3D point cloud matching. Our proposed method is a dual auto-encoder neural network that maps 2D and 3D input into a shared latent space representation. We show that such local cross-domain descriptors in the shared embedding are more discriminative than those obtained from individual training in 2D and 3D domains. To facilitate the training process, we built a new dataset by collecting ≈ 1.4 millions of 2D-3D correspondences with various lighting conditions and settings from publicly available RGB-D scenes. Our descriptor is evaluated in three main experiments: 2D-3D matching, cross-domain retrieval, and sparse-to-dense depth estimation. Experimental results confirm the robustness of our approach as well as its competitive performance not only in solving cross-domain tasks but also in being able to generalize to solve sole 2D and 3D tasks. Our dataset and code are released publicly at https://hkust-vgd.github.io/lcd.
Cite
Text
Pham et al. "LCD: Learned Cross-Domain Descriptors for 2D-3D Matching." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I07.6859Markdown
[Pham et al. "LCD: Learned Cross-Domain Descriptors for 2D-3D Matching." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/pham2020aaai-lcd/) doi:10.1609/AAAI.V34I07.6859BibTeX
@inproceedings{pham2020aaai-lcd,
title = {{LCD: Learned Cross-Domain Descriptors for 2D-3D Matching}},
author = {Pham, Quang-Hieu and Uy, Mikaela Angelina and Hua, Binh-Son and Nguyen, Duc Thanh and Roig, Gemma and Yeung, Sai-Kit},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2020},
pages = {11856-11864},
doi = {10.1609/AAAI.V34I07.6859},
url = {https://mlanthology.org/aaai/2020/pham2020aaai-lcd/}
}