Extreme Rotation Estimation in the Wild

Abstract

We present a technique and benchmark dataset for estimating the relative 3D orientation between a pair of Internet images captured in an extreme setting, where the images have limited or non-overlapping field of views. Prior work targeting extreme rotation estimation assume constrained 3D environments and emulate perspective images by cropping regions from panoramic views. However, real images captured in the wild are highly diverse, exhibiting variation in both appearance and camera intrinsics. In this work, we propose a Transformer-based method for estimating relative rotations in extreme real-world settings, and contribute the ExtremeLandmarkPairs dataset, assembled from scene-level Internet photo collections. Our evaluation demonstrates that our approach succeeds in estimating the relative rotations in a wide variety of extreme-view Internet image pairs, outperforming various baselines, including dedicated rotation estimation techniques and contemporary 3D reconstruction methods.

Cite

Text

Bezalel et al. "Extreme Rotation Estimation in the Wild." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.00107

Markdown

[Bezalel et al. "Extreme Rotation Estimation in the Wild." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/bezalel2025cvpr-extreme/) doi:10.1109/CVPR52734.2025.00107

BibTeX

@inproceedings{bezalel2025cvpr-extreme,
  title     = {{Extreme Rotation Estimation in the Wild}},
  author    = {Bezalel, Hana and Ankri, Dotan and Cai, Ruojin and Averbach-Elor, Hadar},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {1061-1070},
  doi       = {10.1109/CVPR52734.2025.00107},
  url       = {https://mlanthology.org/cvpr/2025/bezalel2025cvpr-extreme/}
}