Coherent Online Video Style Transfer

Abstract

Training a feed-forward network for the fast neural style transfer of images has proven successful, but the naive extension of processing videos frame by frame is prone to producing flickering results. We propose the first end-to-end network for online video style transfer, which generates temporally coherent stylized video sequences in near real-time. Two key ideas include an efficient network by incorporating short-term coherence, and propagating short-term coherence to long-term, which ensures consistency over a longer period of time. Our network can incorporate different image stylization networks and clearly outperforms the per-frame baseline both qualitatively and quantitatively. Moreover, it can achieve visually comparable coherence to optimization-based video style transfer, but is three orders of magnitude faster.

Cite

Text

Chen et al. "Coherent Online Video Style Transfer." International Conference on Computer Vision, 2017. doi:10.1109/ICCV.2017.126

Markdown

[Chen et al. "Coherent Online Video Style Transfer." International Conference on Computer Vision, 2017.](https://mlanthology.org/iccv/2017/chen2017iccv-coherent/) doi:10.1109/ICCV.2017.126

BibTeX

@inproceedings{chen2017iccv-coherent,
  title     = {{Coherent Online Video Style Transfer}},
  author    = {Chen, Dongdong and Liao, Jing and Yuan, Lu and Yu, Nenghai and Hua, Gang},
  booktitle = {International Conference on Computer Vision},
  year      = {2017},
  doi       = {10.1109/ICCV.2017.126},
  url       = {https://mlanthology.org/iccv/2017/chen2017iccv-coherent/}
}