ElliPose: Stereoscopic 3D Human Pose Estimation by Fitting Ellipsoids

Abstract

One of the most relevant tasks for augmented and virtual reality applications is the interaction of virtual objects with real humans which requires accurate 3D human pose predictions. Obtaining accurate 3D human poses requires careful camera calibration which is difficult for non-technical personal or in a pop-up scenario. Recent markerless motion capture approaches require accurate camera calibration at least for the final triangulation step. Instead, we solve this problem by presenting ElliPose, Stereoscopic 3D Human Pose Estimation by Fitting Ellipsoids, where we jointly estimate the 3D human as well as the camera pose. We exploit the fact that bones do not change in length over the course of a sequence and thus their relative trajectories have to lie on the surface of a sphere which we can utilize to iteratively correct the camera and 3D pose estimation. As another use-case we demonstrate that our approach can be used as replacement for ground-truth 3D poses to train monocular 3D pose estimators. We show that our method produces competitive results even when comparing with state-of-the-art methods that use more cameras or ground-truth camera extrinsics.

Cite

Text

Grund et al. "ElliPose: Stereoscopic 3D Human Pose Estimation by Fitting Ellipsoids." Winter Conference on Applications of Computer Vision, 2023.

Markdown

[Grund et al. "ElliPose: Stereoscopic 3D Human Pose Estimation by Fitting Ellipsoids." Winter Conference on Applications of Computer Vision, 2023.](https://mlanthology.org/wacv/2023/grund2023wacv-ellipose/)

BibTeX

@inproceedings{grund2023wacv-ellipose,
  title     = {{ElliPose: Stereoscopic 3D Human Pose Estimation by Fitting Ellipsoids}},
  author    = {Grund, Christian and Tanke, Julian and Gall, Jürgen},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2023},
  pages     = {2871-2881},
  url       = {https://mlanthology.org/wacv/2023/grund2023wacv-ellipose/}
}