Creating Summaries from User Videos
Abstract
This paper proposes a novel approach and a new benchmark for video summarization. Thereby we focus on user videos, which are raw videos containing a set of interesting events. Our method starts by segmenting the video by using a novel “superframe” segmentation, tailored to raw videos. Then, we estimate visual interestingness per superframe using a set of low-, mid- and high-level features. Based on this scoring, we select an optimal subset of superframes to create an informative and interesting summary. The introduced benchmark comes with multiple human created summaries, which were acquired in a controlled psychological experiment. This data paves the way to evaluate summarization methods objectively and to get new insights in video summarization. When evaluating our method, we find that it generates high-quality results, comparable to manual, human-created summaries.
Cite
Text
Gygli et al. "Creating Summaries from User Videos." European Conference on Computer Vision, 2014. doi:10.1007/978-3-319-10584-0_33Markdown
[Gygli et al. "Creating Summaries from User Videos." European Conference on Computer Vision, 2014.](https://mlanthology.org/eccv/2014/gygli2014eccv-creating/) doi:10.1007/978-3-319-10584-0_33BibTeX
@inproceedings{gygli2014eccv-creating,
title = {{Creating Summaries from User Videos}},
author = {Gygli, Michael and Grabner, Helmut and Riemenschneider, Hayko and Van Gool, Luc},
booktitle = {European Conference on Computer Vision},
year = {2014},
pages = {505-520},
doi = {10.1007/978-3-319-10584-0_33},
url = {https://mlanthology.org/eccv/2014/gygli2014eccv-creating/}
}