Learning Intrinsic Video Content Using Levenshtein Distance in Graph Partitioning
Abstract
We present a novel approach for automatically learning models of temporal trajectories extracted from video data. Instead of using a representation of linearly time-normalised vectors of fixed-length, our approach makes use of Dynamic Time Warp distance as a similarity measure to capture the underlying ordered structure of variable-length temporal data while removing the non-linear warping of the time scale. We reformulate the structure learning problem as an optimal graph-partitioning of the dataset to solely exploit Dynamic Time Warp similarity weights without the need for intermediate cluster centroid representations. We extend the graph partitioning method and in particular, the Normalised Cut model originally introduced for static image segmentation to unsupervised clustering of temporal trajectories with fully automated model order selection. By computing hierarchical average Dynamic Time Warp for each cluster, we learn warp-free trajectory models and recover the time warp profiles and structural variance in the data. We demonstrate the approach on modelling trajectories of continuous hand-gestures and moving objects in an indoor environment.
Cite
Text
Ng and Gong. "Learning Intrinsic Video Content Using Levenshtein Distance in Graph Partitioning." European Conference on Computer Vision, 2002. doi:10.1007/3-540-47979-1_45Markdown
[Ng and Gong. "Learning Intrinsic Video Content Using Levenshtein Distance in Graph Partitioning." European Conference on Computer Vision, 2002.](https://mlanthology.org/eccv/2002/ng2002eccv-learning/) doi:10.1007/3-540-47979-1_45BibTeX
@inproceedings{ng2002eccv-learning,
title = {{Learning Intrinsic Video Content Using Levenshtein Distance in Graph Partitioning}},
author = {Ng, Jeffrey and Gong, Shaogang},
booktitle = {European Conference on Computer Vision},
year = {2002},
pages = {670-684},
doi = {10.1007/3-540-47979-1_45},
url = {https://mlanthology.org/eccv/2002/ng2002eccv-learning/}
}