Unsupervised Learning of Skeletons from Motion

Abstract

Humans demonstrate a remarkable ability to parse complicated motion sequences into their constituent structures and motions. We investigate this problem, attempting to learn the structure of one or more articulated objects, given a time-series of two-dimensional feature positions. We model the observed sequence in terms of “stick figure” objects, under the assumption that the relative joint angles between sticks can change over time, but their lengths and connectivities are fixed. We formulate the problem in a single probabilistic model that includes multiple sub-components: associating the features with particular sticks, determining the proper number of sticks, and finding which sticks are physically joined. We test the algorithm on challenging datasets of 2D projections of optical human motion capture and feature trajectories from real videos.

Cite

Text

Ross et al. "Unsupervised Learning of Skeletons from Motion." European Conference on Computer Vision, 2008. doi:10.1007/978-3-540-88690-7_42

Markdown

[Ross et al. "Unsupervised Learning of Skeletons from Motion." European Conference on Computer Vision, 2008.](https://mlanthology.org/eccv/2008/ross2008eccv-unsupervised/) doi:10.1007/978-3-540-88690-7_42

BibTeX

@inproceedings{ross2008eccv-unsupervised,
  title     = {{Unsupervised Learning of Skeletons from Motion}},
  author    = {Ross, David A. and Tarlow, Daniel and Zemel, Richard S.},
  booktitle = {European Conference on Computer Vision},
  year      = {2008},
  pages     = {560-573},
  doi       = {10.1007/978-3-540-88690-7_42},
  url       = {https://mlanthology.org/eccv/2008/ross2008eccv-unsupervised/}
}