Leaping from 2D Detection to Efficient 6DoF Object Pose Estimation
Abstract
Estimating 6DoF object poses from single RGB images is very challenging due to severe occlusions and large search space of camera poses. Keypoint voting based methods have demonstrated its effectiveness and superiority on predicting object poses. However, those approaches are often affected by inaccurate semantic segmentation in computing the keypoint locations. To enable our model to focus on local regions without being distracted by backgrounds, we first localize object regions by a 2D object detector. In doing so, we not only reduce the search space of keypoints but also improve the robustness of the pose estimation. Moreover, since symmetric objects may suffer ambiguity along the symmetric dimension, we propose to select keypoints on the geometrically symmetric locations to resolve the ambiguity. The extensive experimental results on seven different datasets of the BOP challenge benchmark demonstrate that our method outperforms the state-of-the-art and achieves the 3-rd place in the BOP challenge.
Cite
Text
Liu et al. "Leaping from 2D Detection to Efficient 6DoF Object Pose Estimation." European Conference on Computer Vision Workshops, 2020. doi:10.1007/978-3-030-66096-3_47Markdown
[Liu et al. "Leaping from 2D Detection to Efficient 6DoF Object Pose Estimation." European Conference on Computer Vision Workshops, 2020.](https://mlanthology.org/eccvw/2020/liu2020eccvw-leaping/) doi:10.1007/978-3-030-66096-3_47BibTeX
@inproceedings{liu2020eccvw-leaping,
title = {{Leaping from 2D Detection to Efficient 6DoF Object Pose Estimation}},
author = {Liu, Jinhui and Zou, Zhikang and Ye, Xiaoqing and Tan, Xiao and Ding, Errui and Xu, Feng and Yu, Xin},
booktitle = {European Conference on Computer Vision Workshops},
year = {2020},
pages = {707-714},
doi = {10.1007/978-3-030-66096-3_47},
url = {https://mlanthology.org/eccvw/2020/liu2020eccvw-leaping/}
}