Joint Event Detection and Description in Continuous Video Streams
Abstract
Dense video captioning involves first localizing events in a video and then generating captions for the identified events. We present the Joint Event Detection and Description Network (JEDDi-Net) for solving this task in an end-to-end fashion, which encodes the input video stream with three-dimensional convolutional layers, proposes variable-length temporal events based on pooled features, and then uses a two-level hierarchical LSTM module with context modeling to transcribe the event proposals into captions. We show the effectiveness of our proposed JEDDi-Net on the large-scale ActivityNet Captions dataset.
Cite
Text
Xu et al. "Joint Event Detection and Description in Continuous Video Streams." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2019. doi:10.1109/WACVW.2019.00011Markdown
[Xu et al. "Joint Event Detection and Description in Continuous Video Streams." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2019.](https://mlanthology.org/wacvw/2019/xu2019wacvw-joint/) doi:10.1109/WACVW.2019.00011BibTeX
@inproceedings{xu2019wacvw-joint,
title = {{Joint Event Detection and Description in Continuous Video Streams}},
author = {Xu, Huijuan and Li, Boyang and Ramanishka, Vasili and Sigal, Leonid and Saenko, Kate},
booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision Workshops},
year = {2019},
pages = {25-26},
doi = {10.1109/WACVW.2019.00011},
url = {https://mlanthology.org/wacvw/2019/xu2019wacvw-joint/}
}