Unsupervised Activity Perception by Hierarchical Bayesian Models

Abstract

We propose a novel unsupervised learning framework for activity perception. To understand activities in complicated scenes from visual data, we propose a hierarchical Bayesian model to connect three elements: low-level visual features, simple "atomic" activities, and multi-agent interactions. Atomic activities are modeled as distributions over low-level visual features, and interactions are modeled as distributions over atomic activities. Our models improve existing language models such as Latent Dirichlet Allocation (LDA) and Hierarchical Dirichlet Process (HDP) by modeling interactions without supervision. Our data sets are challenging video sequences from crowded traffic scenes with many kinds of activities co-occurring. Our approach provides a summary of typical atomic activities and interactions in the scene. Unusual activities and interactions are found, with natural probabilistic explanations. Our method supports flexible high-level queries on activities and interactions using atomic activities as components.

Cite

Text

Wang et al. "Unsupervised Activity Perception by Hierarchical Bayesian Models." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007. doi:10.1109/CVPR.2007.383072

Markdown

[Wang et al. "Unsupervised Activity Perception by Hierarchical Bayesian Models." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2007.](https://mlanthology.org/cvpr/2007/wang2007cvpr-unsupervised/) doi:10.1109/CVPR.2007.383072

BibTeX

@inproceedings{wang2007cvpr-unsupervised,
  title     = {{Unsupervised Activity Perception by Hierarchical Bayesian Models}},
  author    = {Wang, Xiaogang and Ma, Xiaoxu and Grimson, Eric},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2007},
  doi       = {10.1109/CVPR.2007.383072},
  url       = {https://mlanthology.org/cvpr/2007/wang2007cvpr-unsupervised/}
}