Creating Summaries from User Videos

Abstract

This paper proposes a novel approach and a new benchmark for video summarization. Thereby we focus on user videos, which are raw videos containing a set of interesting events. Our method starts by segmenting the video by using a novel “superframe” segmentation, tailored to raw videos. Then, we estimate visual interestingness per superframe using a set of low-, mid- and high-level features. Based on this scoring, we select an optimal subset of superframes to create an informative and interesting summary. The introduced benchmark comes with multiple human created summaries, which were acquired in a controlled psychological experiment. This data paves the way to evaluate summarization methods objectively and to get new insights in video summarization. When evaluating our method, we find that it generates high-quality results, comparable to manual, human-created summaries.

Cite

Text

Gygli et al. "Creating Summaries from User Videos." European Conference on Computer Vision, 2014. doi:10.1007/978-3-319-10584-0_33

Markdown

[Gygli et al. "Creating Summaries from User Videos." European Conference on Computer Vision, 2014.](https://mlanthology.org/eccv/2014/gygli2014eccv-creating/) doi:10.1007/978-3-319-10584-0_33

BibTeX

@inproceedings{gygli2014eccv-creating,
  title     = {{Creating Summaries from User Videos}},
  author    = {Gygli, Michael and Grabner, Helmut and Riemenschneider, Hayko and Van Gool, Luc},
  booktitle = {European Conference on Computer Vision},
  year      = {2014},
  pages     = {505-520},
  doi       = {10.1007/978-3-319-10584-0_33},
  url       = {https://mlanthology.org/eccv/2014/gygli2014eccv-creating/}
}