H-Net: Neural Network for Cross-Domain Image Patch Matching
Abstract
Describing the same scene with different imaging style or rendering image from its 3D model gives us different domain images. Different domain images tend to have a gap and different local appearances, which raise the main challenge on the cross-domain image patch matching. In this paper, we propose to incorporate AutoEncoder into the Siamese network, named as H-Net, of which the structural shape resembles the letter H. The H-Net achieves state-of-the-art performance on the cross-domain image patch matching. Furthermore, we improved H-Net to H-Net++. The H-Net++ extracts invariant feature descriptors in cross-domain image patches and achieves state-of-the-art performance by feature retrieval in Euclidean space. As there is no benchmark dataset including cross-domain images, we made a cross-domain image dataset which consists of camera images, rendering images from UAV 3D model, and images generated by CycleGAN algorithm. Experiments show that the proposed H-Net and H-Net++ outperform the existing algorithms. Our code and cross-domain image dataset are available at https://github.com/Xylon-Sean/H-Net.
Cite
Text
Liu et al. "H-Net: Neural Network for Cross-Domain Image Patch Matching." International Joint Conference on Artificial Intelligence, 2018. doi:10.24963/IJCAI.2018/119Markdown
[Liu et al. "H-Net: Neural Network for Cross-Domain Image Patch Matching." International Joint Conference on Artificial Intelligence, 2018.](https://mlanthology.org/ijcai/2018/liu2018ijcai-h/) doi:10.24963/IJCAI.2018/119BibTeX
@inproceedings{liu2018ijcai-h,
title = {{H-Net: Neural Network for Cross-Domain Image Patch Matching}},
author = {Liu, Weiquan and Shen, Xuelun and Wang, Cheng and Zhang, Zhihong and Wen, Chenglu and Li, Jonathan},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2018},
pages = {856-863},
doi = {10.24963/IJCAI.2018/119},
url = {https://mlanthology.org/ijcai/2018/liu2018ijcai-h/}
}