Cross-Identity Video Motion Retargeting with Joint Transformation and Synthesis
Abstract
In this paper, we propose a novel dual-branch Transformation-Synthesis network (TS-Net), for video motion retargeting. Given one subject video and one driving video, TS-Net can produce a new plausible video with the subject appearance of the subject video and motion pattern of the driving video. TS-Net consists of a warp-based transformation branch and a warp-free synthesis branch. The novel design of dual branches combines the strengths of deformation-grid-based transformation and warp-free generation for better identity preservation and robustness to occlusion in the synthesized videos. A mask-aware similarity module is further introduced to the transformation branch to reduce computational overhead. Experimental results on face and dance datasets show that TS-Net achieves better performance in video motion retargeting than several state-of-the-art models as well as its single-branch variants. Our code is available at https://github.com/nihaomiao/WACV23_TSNet.
Cite
Text
Ni et al. "Cross-Identity Video Motion Retargeting with Joint Transformation and Synthesis." Winter Conference on Applications of Computer Vision, 2023.Markdown
[Ni et al. "Cross-Identity Video Motion Retargeting with Joint Transformation and Synthesis." Winter Conference on Applications of Computer Vision, 2023.](https://mlanthology.org/wacv/2023/ni2023wacv-crossidentity/)BibTeX
@inproceedings{ni2023wacv-crossidentity,
title = {{Cross-Identity Video Motion Retargeting with Joint Transformation and Synthesis}},
author = {Ni, Haomiao and Liu, Yihao and Huang, Sharon X. and Xue, Yuan},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2023},
pages = {412-422},
url = {https://mlanthology.org/wacv/2023/ni2023wacv-crossidentity/}
}