Image as an IMU: Estimating Camera Motion from a Single Motion-Blurred Image
Abstract
In many robotics and VR/AR applications, fast camera motions lead to a high level of motion blur, causing existing camera pose estimation methods to fail. In this work, we propose a novel framework that leverages motion blur as a rich cue for motion estimation rather than treating it as an unwanted artifact. Our approach works by predicting a dense motion flow field and a monocular depth map directly from a single motion-blurred image. We then recover the instantaneous camera velocity by solving a linear least squares problem under the small motion assumption. In essence, our method produces an IMU-like measurement that robustly captures fast and aggressive camera movements. To train our model, we construct a large-scale dataset with realistic synthetic motion blur derived from ScanNet++v2 and further refine our model by training end-to-end on real data using our fully differentiable pipeline. Extensive evaluations on real-world benchmarks demonstrate that our method achieves state-of-the-art angular and translational velocity estimates, outperforming current methods like MASt3R and COLMAP.
Cite
Text
Chen and Clark. "Image as an IMU: Estimating Camera Motion from a Single Motion-Blurred Image." International Conference on Computer Vision, 2025.Markdown
[Chen and Clark. "Image as an IMU: Estimating Camera Motion from a Single Motion-Blurred Image." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/chen2025iccv-image/)BibTeX
@inproceedings{chen2025iccv-image,
title = {{Image as an IMU: Estimating Camera Motion from a Single Motion-Blurred Image}},
author = {Chen, Jerred and Clark, Ronald},
booktitle = {International Conference on Computer Vision},
year = {2025},
pages = {90-99},
url = {https://mlanthology.org/iccv/2025/chen2025iccv-image/}
}