Temporally-Dependent Dirichlet Process Mixtures for Egocentric Video Segmentation
Abstract
In this paper, we present a novel approach for segmenting video into large regions of generally similar activity. Based on the Dirichlet Process Multinomial Mixture model, we introduce temporal dependency into the inference algorithm, allowing our method to automatically create long segments with high saliency while ignoring small, inconsequential interruptions. We evaluate our algorithm and other topic models with both synthetic datasets and real-world video. Additionally, applicability to image segmentation is shown. Results show that our method outperforms related methods with respect to accuracy and noise removal.
Cite
Text
Barker and Davis. "Temporally-Dependent Dirichlet Process Mixtures for Egocentric Video Segmentation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2014. doi:10.1109/CVPRW.2014.88Markdown
[Barker and Davis. "Temporally-Dependent Dirichlet Process Mixtures for Egocentric Video Segmentation." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2014.](https://mlanthology.org/cvprw/2014/barker2014cvprw-temporallydependent/) doi:10.1109/CVPRW.2014.88BibTeX
@inproceedings{barker2014cvprw-temporallydependent,
title = {{Temporally-Dependent Dirichlet Process Mixtures for Egocentric Video Segmentation}},
author = {Barker, Joseph W. and Davis, James W.},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
year = {2014},
pages = {571-578},
doi = {10.1109/CVPRW.2014.88},
url = {https://mlanthology.org/cvprw/2014/barker2014cvprw-temporallydependent/}
}