Co-Op: Correspondence-Based Novel Object Pose Estimation
Abstract
We propose Co-op, a novel method for accurately and robustly estimating the 6DoF pose of objects unseen during training from a single RGB image. Our method requires only the CAD model of the target object and can precisely estimate its pose without any additional fine-tuning. While existing model-based methods suffer from inefficiency due to using a large number of templates, our method enables fast and accurate estimation with a small number of templates. This improvement is achieved by finding semi-dense correspondences between the input image and the pre-rendered templates. Our method achieves strong generalization performance by leveraging a hybrid representation that combines patch-level classification and offset regression. Additionally, our pose refinement model estimates probabilistic flow between the input image and the rendered image, refining the initial estimate to an accurate pose using a differentiable PnP layer. We demonstrate that our method not only estimates object poses rapidly but also outperforms existing methods by a large margin on the seven core datasets of the BOP Challenge, achieving state-of-the-art accuracy.
Cite
Text
Moon et al. "Co-Op: Correspondence-Based Novel Object Pose Estimation." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.01085Markdown
[Moon et al. "Co-Op: Correspondence-Based Novel Object Pose Estimation." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/moon2025cvpr-coop/) doi:10.1109/CVPR52734.2025.01085BibTeX
@inproceedings{moon2025cvpr-coop,
title = {{Co-Op: Correspondence-Based Novel Object Pose Estimation}},
author = {Moon, Sungphill and Son, Hyeontae and Hur, Dongcheol and Kim, Sangwook},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2025},
pages = {11622-11632},
doi = {10.1109/CVPR52734.2025.01085},
url = {https://mlanthology.org/cvpr/2025/moon2025cvpr-coop/}
}