Human from Blur: Human Pose Tracking from Blurry Images

Abstract

We propose a method to estimate 3D human poses from substantially blurred images. The key idea is to tackle the inverse problem of image deblurring by modeling the forward problem with a 3D human model, a texture map, and a sequence of poses to describe human motion. The blurring process is then modeled by a temporal image aggregation step. Using a differentiable renderer, we can solve the inverse problem by backpropagating the pixel-wise reprojection error to recover the best human motion representation that explains a single or multiple input images. Since the image reconstruction loss alone is insufficient, we present additional regularization terms. To the best of our knowledge, we present the first method to tackle this problem. Our method consistently outperforms other methods on significantly blurry inputs since they lack one or multiple key functionalities that our method unifies, i.e. image deblurring with sub-frame accuracy and explicit 3D modeling of non-rigid human motion.

Cite

Text

Zhao et al. "Human from Blur: Human Pose Tracking from Blurry Images." International Conference on Computer Vision, 2023. doi:10.1109/ICCV51070.2023.01369

Markdown

[Zhao et al. "Human from Blur: Human Pose Tracking from Blurry Images." International Conference on Computer Vision, 2023.](https://mlanthology.org/iccv/2023/zhao2023iccv-human/) doi:10.1109/ICCV51070.2023.01369

BibTeX

@inproceedings{zhao2023iccv-human,
  title     = {{Human from Blur: Human Pose Tracking from Blurry Images}},
  author    = {Zhao, Yiming and Rozumnyi, Denys and Song, Jie and Hilliges, Otmar and Pollefeys, Marc and Oswald, Martin R.},
  booktitle = {International Conference on Computer Vision},
  year      = {2023},
  pages     = {14905-14915},
  doi       = {10.1109/ICCV51070.2023.01369},
  url       = {https://mlanthology.org/iccv/2023/zhao2023iccv-human/}
}